Shortly after voting to move forward with a pair of subpoenas, the Senate Judiciary Committee has reached an agreement that will see the CEOs of two major social platforms testify voluntarily in November. The hearing will be the second major congressional appearance by tech CEOs arranged this month.
Twitter’s Jack Dorsey and Facebook’s Mark Zuckerberg will answer questions at the hearing, set for November 17 — two weeks after election day. The Republican-led committee is chaired by South Carolina Senator Lindsey Graham, who set the agenda to include the “platforms’ censorship and suppression of New York Post articles.”
According to a new press release from the committee, lawmakers also plan to use the proceedings as a high-profile port-mortem on how Twitter and Facebook fared on and after election day — an issue that lawmakers on both sides will undoubtedly be happy to dig into.
Republicans are eager to press the tech CEOs on how their respective platforms handled a dubious story from the New York Post purporting to report on hacked materials from presidential candidate Joe Biden’s son, Hunter Biden. They view the incident as evidence of their ongoing claims of anti-conservative political bias in platform policy decisions.
While Republicans on the Senate committee led the decision to pressure Zuckerberg and Dorsey into testifying, the committee’s Democrats, who sat out the vote on the subpoenas, will likely bring to the table their own questions about content moderation, as well.
In ye olden days of piracy, RIAA takedown notices were a common thing — I received a few myself. But that’s mostly fallen off as tracking pirates has gotten more difficult. But the RIAA can still issue nastygrams — to the creators of software that could potentially be used to violate copyright, like YouTube downloaders.
One such popular tool used by many developers, YouTube-DL, has been removed from GitHub for the present after an RIAA threat, as noted by Freedom of the Press Foundation’s Parker Higgins earlier today.
This is a different kind of takedown notice than the ones we all remember from the early 2000s, though. Those were the innumerable DMCA notices that said “your website is hosting such-and-such protected content, please take it down.” And they still exist, of course, but lots of that has become automated, with sites like YouTube removing infringing videos before they even go public.
What the RIAA has done here is demand that YouTube -DL be taken down because it violates Section 1201 of U.S. copyright law, which basically bans stuff that gets around DRM. “No person shall circumvent a technological measure that effectively controls access to a work protected under this title.”
That’s so it’s illegal not just to distribute, say, a bootleg Blu-ray disc, but also to break its protections and duplicate it in the first place.
If you stretch that logic a bit, you end up including things like YouTube-DL, which is a command-line tool that takes in a YouTube URL and points the user to the raw video and audio, which of course have to be stored on a server somewhere. With the location of the file that would normally be streamed in the YouTube web player, the user can download a video for offline use or backup.
But what if someone were to use that tool to download the official music video for Taylor Swift’s “Shake it off”? Shock! Horror! Piracy! YouTube-DL enables this, so it must be taken down, they write.
As usual, it only takes a moment to arrive at analogous (or analog) situations that the RIAA has long given up on. For instance, wouldn’t using a screen and audio capture utility accomplish the same thing? What about a camcorder? Or for that matter, a cassette recorder? They’re all used to “circumvent” the DRM placed on Tay’s video by creating an offline copy without the rights-holder’s permission.
Naturally this takedown will do almost nothing to prevent the software, which was probably downloaded and forked thousands of times already, from being used or updated. There are also dozens of sites and apps that do this — and the RIAA by the logic in this letter may very well take action against them as well.
Of course, the RIAA is bound by duty to protect against infringement, and one can’t expect it to stand by idly as people scrape official YouTube accounts to get high-quality bootlegs of artists’ entire discographies. But going after the basic tools is like the old, ineffective “Home taping is killing the music industry” line. No one’s buying it. And if we’re going to talk about wholesale theft of artists, perhaps the RIAA should get its own house in order first — streaming services are paying out pennies with the Association’s blessing. (Go buy stuff on Bandcamp instead.)
Tools like YouTube-DL, like cassette tapes, cameras and hammers, are tech that can be used legally or illegally. Fair use doctrines allow tools like these for good-faith efforts like archiving content that might be lost because Google stops caring, or for people who for one reason or another want to have a local copy of some widely available, free piece of media for personal use.
YouTube and other platforms, likewise in good faith, do what they can to make obvious and large-scale infringement difficult. There’s no “download” button next to the latest Top 40 hit, but there are links to buy it, and if I used a copy — even one I’d bought — as background for my own video, I wouldn’t even be able to put it on YouTube in the first place.
Temporarily removing YouTube-DL’s code from GitHub is a short-sighted reaction to a problem that can’t possibly amount to more than a rounding error in the scheme of things. They probably lose more money to people sharing logins. It or something very much like it will be back soon, a little smarter and a little better, making the RIAA’s job that much harder, and the cycle will repeat.
Maybe the creators of Whack-a-Mole will sue the RIAA for infringement on their unique IP.
A California court weighs in as Prop. 22 looms, Google removes popular apps over data collection practices and the Senate subpoenas Jack Dorsey and Mark Zuckerberg. This is your Daily Crunch for October 23, 2020.
The big story: Uber and Lyft defeated again in court
A California appeals court ruled that yes, a new state law applies to Uber and Lyft drivers, meaning that they must be classified as employees, rather than independent contractors. The judge ruled that contrary to the rideshare companies’ arguments, any financial harm does not “rise to the level of irreparable harm.”
However, the decision will not take effect for 30 days — suggesting that the real determining factor will be Proposition 22, a statewide ballot measure backed by Uber and Lyft that would keep drivers as contractors while guaranteeing things like minimum compensation and healthcare subsidies.
“This ruling makes it more urgent than ever for voters to stand with drivers and vote yes on Prop. 22,” a Lyft spokesperson told TechCrunch.
The tech giants
Google removes 3 Android apps for children, with 20M+ downloads between them, over data collection violations — Researchers at the International Digital Accountability Council found that a trio of popular and seemingly innocent-looking apps aimed at younger users were violating Google’s data collection policies.
Huawei reports slowing growth as its operations ‘face significant challenges’ — The full impact of U.S. trade restrictions hasn’t been realized yet, because the government has granted Huawei several waivers.
Senate subpoenas could force Zuckerberg and Dorsey to testify on New York Post controversy — The Senate Judiciary Committee voted in favor of issuing subpoenas for Facebook’s Mark Zuckerberg and Twitter’s Jack Dorsey.
Startups, funding and venture capital
Quibi says it will shut down in early December — A newly published support page on the Quibi site says streaming will end “on or about December 1, 2020.”
mmhmm, Phil Libin’s new startup, acquires Memix to add enhanced filters to its video presentation toolkit — Memix has built a series of filters you can apply to videos to change the lighting, the details in the background or across the whole screen.
Nordic challenger bank Lunar raises €40M Series C, plans to enter the ‘buy now, pay later’ space — Lunar started out as a personal finance manager app but acquired a full banking license in 2019.
Advice and analysis from Extra Crunch
Here’s how fast a few dozen startups grew in Q3 2020 — This is as close to private company earnings reports as we can manage.
The short, strange life of Quibi — Everything you need to know about the Quibi story, all in one place.
(Reminder: Extra Crunch is our membership program, which aims to democratize information about startups. You can sign up here.)
Everything else
France rebrands contact-tracing app in an effort to boost downloads — France’s contact-tracing app has been updated and is now called TousAntiCovid, which means “everyone against Covid.”
Representatives propose bill limiting presidential internet ‘kill switch’ — The bill would limit the president’s ability to shut down the internet at will.
The Daily Crunch is TechCrunch’s roundup of our biggest and most important stories. If you’d like to get this delivered to your inbox every day at around 3pm Pacific, you can subscribe here.
Research papers come out far too rapidly for anyone to read them all, especially in the field of machine learning, which now affects (and produces papers in) practically every industry and company. This column aims to collect the most relevant recent discoveries and papers — particularly in but not limited to artificial intelligence — and explain why they matter.
This week, a startup that’s using UAV drones for mapping forests, a look at how machine learning can map social media networks and predict Alzheimer’s, improving computer vision for space-based sensors and other news regarding recent technological advances.
Predicting Alzheimer’s through speech patterns
Machine learning tools are being used to aid diagnosis in many ways, since they’re sensitive to patterns that humans find difficult to detect. IBM researchers have potentially found such patterns in speech that are predictive of the speaker developing Alzheimer’s disease.
The system only needs a couple minutes of ordinary speech in a clinical setting. The team used a large set of data (the Framingham Heart Study) going back to 1948, allowing patterns of speech to be identified in people who would later develop Alzheimer’s. The accuracy rate is about 71% or 0.74 area under the curve for those of you more statistically informed. That’s far from a sure thing, but current basic tests are barely better than a coin flip in predicting the disease this far ahead of time.
This is very important because the earlier Alzheimer’s can be detected, the better it can be managed. There’s no cure, but there are promising treatments and practices that can delay or mitigate the worst symptoms. A non-invasive, quick test of well people like this one could be a powerful new screening tool and is also, of course, an excellent demonstration of the usefulness of this field of tech.
(Don’t read the paper expecting to find exact symptoms or anything like that — the array of speech features aren’t really the kind of thing you can look out for in everyday life.)
So-cell networks
Making sure your deep learning network generalizes to data outside its training environment is a key part of any serious ML research. But few attempt to set a model loose on data that’s completely foreign to it. Perhaps they should!
Researchers from Uppsala University in Sweden took a model used to identify groups and connections in social media, and applied it (not unmodified, of course) to tissue scans. The tissue had been treated so that the resultant images produced thousands of tiny dots representing mRNA.
Normally the different groups of cells, representing types and areas of tissue, would need to be manually identified and labeled. But the graph neural network, created to identify social groups based on similarities like common interests in a virtual space, proved it could perform a similar task on cells. (See the image at top.)
“We’re using the latest AI methods — specifically, graph neural networks, developed to analyze social networks — and adapting them to understand biological patterns and successive variation in tissue samples. The cells are comparable to social groupings that can be defined according to the activities they share in their social networks,” said Uppsala’s Carolina Wählby.
It’s an interesting illustration not just of the flexibility of neural networks, but of how structures and architectures repeat at all scales and in all contexts. As without, so within, if you will.
Drones in nature
The vast forests of our national parks and timber farms have countless trees, but you can’t put “countless” on the paperwork. Someone has to make an actual estimate of how well various regions are growing, the density and types of trees, the range of disease or wildfire, and so on. This process is only partly automated, as aerial photography and scans only reveal so much, while on-the-ground observation is detailed but extremely slow and limited.
Treeswift aims to take a middle path by equipping drones with the sensors they need to both navigate and accurately measure the forest. By flying through much faster than a walking person, they can count trees, watch for problems and generally collect a ton of useful data. The company is still very early-stage, having spun out of the University of Pennsylvania and acquired an SBIR grant from the NSF.
“Companies are looking more and more to forest resources to combat climate change but you don’t have a supply of people who are growing to meet that need,” Steven Chen, co-founder and CEO of Treeswift and a doctoral student in Computer and Information Science (CIS) at Penn Engineering said in a Penn news story. “I want to help make each forester do what they do with greater efficiency. These robots will not replace human jobs. Instead, they’re providing new tools to the people who have the insight and the passion to manage our forests.”
Another area where drones are making lots of interesting moves is underwater. Oceangoing autonomous submersibles are helping map the sea floor, track ice shelves and follow whales. But they all have a bit of an Achilles’ heel in that they need to periodically be picked up, charged and their data retrieved.
Purdue engineering professor Nina Mahmoudian has created a docking system by which submersibles can easily and automatically connect for power and data exchange.

A yellow marine robot (left, underwater) finds its way to a mobile docking station to recharge and upload data before continuing a task. (Purdue University photo/Jared Pike)
The craft needs a special nosecone, which can find and plug into a station that establishes a safe connection. The station can be an autonomous watercraft itself, or a permanent feature somewhere — what matters is that the smaller craft can make a pit stop to recharge and debrief before moving on. If it’s lost (a real danger at sea), its data won’t be lost with it.
You can see the setup in action below:
https://youtu.be/kS0-qc_r0
Sound in theory
Drones may soon become fixtures of city life as well, though we’re probably some ways from the automated private helicopters some seem to think are just around the corner. But living under a drone highway means constant noise — so people are always looking for ways to reduce turbulence and resultant sound from wings and propellers.
Researchers at the King Abdullah University of Science and Technology found a new, more efficient way to simulate the airflow in these situations; fluid dynamics is essentially as complex as you make it, so the trick is to apply your computing power to the right parts of the problem. They were able to render only flow near the surface of the theoretical aircraft in high resolution, finding past a certain distance there was little point knowing exactly what was happening. Improvements to models of reality don’t always need to be better in every way — after all, the results are what matter.
Machine learning in space
Computer vision algorithms have come a long way, and as their efficiency improves they are beginning to be deployed at the edge rather than at data centers. In fact it’s become fairly common for camera-bearing objects like phones and IoT devices to do some local ML work on the image. But in space it’s another story.
Performing ML work in space was until fairly recently simply too expensive power-wise to even consider. That’s power that could be used to capture another image, transmit the data to the surface, etc. HyperScout 2 is exploring the possibility of ML work in space, and its satellite has begun applying computer vision techniques immediately to the images it collects before sending them down. (“Here’s a cloud — here’s Portugal — here’s a volcano…”)
For now there’s little practical benefit, but object detection can be combined with other functions easily to create new use cases, from saving power when no objects of interest are present, to passing metadata to other tools that may work better if informed.
In with the old, out with the new
Machine learning models are great at making educated guesses, and in disciplines where there’s a large backlog of unsorted or poorly documented data, it can be very useful to let an AI make a first pass so that graduate students can use their time more productively. The Library of Congress is doing it with old newspapers, and now Carnegie Mellon University’s libraries are getting into the spirit.
CMU’s million-item photo archive is in the process of being digitized, but to make it useful to historians and curious browsers it needs to be organized and tagged — so computer vision algorithms are being put to work grouping similar images, identifying objects and locations, and doing other valuable basic cataloguing tasks.
“Even a partly successful project would greatly improve the collection metadata, and could provide a possible solution for metadata generation if the archives were ever funded to digitize the entire collection,” said CMU’s Matt Lincoln.
A very different project, yet one that seems somehow connected, is this work by a student at the Escola Politécnica da Universidade de Pernambuco in Brazil, who had the bright idea to try sprucing up some old maps with machine learning.
The tool they used takes old line-drawing maps and attempts to create a sort of satellite image based on them using a Generative Adversarial Network; GANs essentially attempt to trick themselves into creating content they can’t tell apart from the real thing.
Well, the results aren’t what you might call completely convincing, but it’s still promising. Such maps are rarely accurate but that doesn’t mean they’re completely abstract — recreating them in the context of modern mapping techniques is a fun idea that might help these locations seem less distant.
This year has shaken up venture capital, turning a hot early start to 2020 into a glacial period permeated with fear during the early days of COVID-19. That ice quickly melted as venture capitalists discovered that demand for software and other services that startups provide was accelerating, pushing many young tech companies back into growth mode, and investors back into the check-writing arena.
Boston has been an exemplar of the trend, with early pandemic caution dissolving into rapid-fire dealmaking as summer rolled into fall.
We collated new data that underscores the trend, showing that Boston’s third quarter looks very solid compared to its peer groups, and leads greater New England’s share of American venture capital higher during the three-month period.
For our October look at Boston and its startup scene, let’s get into the data and then understand how a new cohort of founders is cropping up among the city’s educational network.
A strong Q3, a strong 2020
Boston’s third quarter was strong, effectively matching the capital raised in New York City during the three-month period. As we head into the fourth quarter, it appears that the silver medal in American startup ecosystems is up for grabs based on what happens in Q4.
Boston could start 2021 as the number-two place to raise venture capital in the country. Or New York City could pip it at the finish line. Let’s check the numbers.
According to PitchBook data shared with TechCrunch, the metro Boston area raised $4.34 billion in venture capital during the third quarter. New York City and its metro area managed $4.45 billion during the same time period, an effective tie. Los Angeles and its own metro area managed just $3.90 billion.
In 2020 the numbers tilt in Boston’s favor, with the city and surrounding area collecting $12.83 billion in venture capital. New York City came in second through Q3, with $12.30 billion in venture capital. Los Angeles was a distant third at $8.66 billion for the year through Q3.