‘Metropolis’, Sherlock Holmes Finally Enter the Public Domain 95 Years Later

Guess what’s finally entering America’s public domain today? Appropriately enough, it’s Marcel Proust’s 1927 novel Remembrance of Things Past.

Also entering the public domain today are thousands of other books, plus the music and lyrics of hundreds of songs, and even several silent movies.

Fritz Lang’s sci-fi classic Metropolis enters the public domain today — and so does the Laurel & Hardy comedy Battle of the Century (which culminates with one of Hollywod’s first pie fights), according to Duke University’s Center for the Study of the Public Domain:
This is actually the second time that Metropolis has gone into the US public domain. The first was in 1955, when its initial 28-year term expired and the rights holders did not renew the copyright. Then in 1996 a new law restored the copyrights in qualifying foreign works. Metropolis, along with thousands of other works, was pulled out of the public domain, and now reenters it after the expiration of the 95-year term, with the once missing scenes available for anyone to reuse.
They also note that some material is in the public domain from the beginning, including government works like the images from the James Webb telescope.

But for other works, today is a big and important day, writes the Associated Press:
Alongside the short-story collection “The Case-Book of Sherlock Holmes,” books such as Virginia Woolf’s “To The Lighthouse,” Ernest Hemingway’s “Men Without Women,” William Faulkner’s “Mosquitoes” and Agatha Christie’s “The Big Four” — an Hercule Poirot mystery — will become public domain as the calendar turns to 2023. Once a work enters the public domain it can legally be shared, performed, reused, repurposed or sampled without permission or cost.

The works from 1927 were originally supposed to be copyrighted for 75 years, but the 1998 Copyright Term Extension Act delayed opening them up for an additional 20 years. While many prominent works on the list used those extra two decades to earn their copyright holders good money, a Duke University expert says the copyright protections also applied to “all of the works whose commercial viability had long subsided.”

“For the vast majority — probably 99% — of works from 1927, no copyright holder financially benefited from continued copyright. Yet they remained off limits, for no good reason,” Jennifer Jenkins, director of Duke’s Center for the Study of the Public Domain, wrote in a blog post heralding “Public Domain Day 2023.” That long U.S. copyright period meant many works that would now become available have long since been lost, because they were not profitable to maintain by the legal owners, but couldn’t be used by others. On the Duke list are such “lost” films like Victor Fleming’s “The Way of All Flesh” and Tod Browning’s “London After Midnight….”

Also entering the public domain today:

– Willa Cather’s Death Comes for the Archbishop
– A. A. Milne’s Now We Are Six (illustrations by E. H. Shepard)
– Franklin W. Dixon’s The Tower Treasure — the first Hardy Boys book
– Herman Hesse’s Steppenwolf (German version)
– The song “My Blue Heaven”
– Songs by Duke Ellington and Louis Armstrong
– Alfred Hitchcock’s early silent movie The Lodger

The UK-based newspaper the Observer adds:
For those readers who do not reside in the US, there is perhaps another reason for celebrating today, because copyright terms are longer in the US than they are in other parts of the world, including the EU and the UK. And therein lies a story about intellectual property laws and the power of political lobbying in a so-called liberal democracy…. The term was gradually lengthened in small increments by Congress until 1976, when it was extended by 19 years to 75 years and then in 1998 by the Sonny Bono Act. So, as the legal scholar Lawrence Lessig puts it, “in the 20 years after the Sonny Bono Act, while 1 million patents will pass into the public domain, zero copyrights will pass into the public domain by virtue of the expiration of a copyright term”….

[T]he end result is that American citizens have had to wait two decades to be free to adapt and reuse works to which we Europeans have had easy access….

The issue highlighted by Public Domain Day is not that intellectual property is evil but that aspects of it — especially copyright — have been monopolised and weaponised by corporate interests and that legislators have been supine in the face of their lobbying. Authors and inventors need protection against being ripped off. It’s obviously important that clever people are rewarded for their creativity and the patent system does that quite well. But if a patent only lasts for 20 years, why on earth should copyright last for life plus 70 years for a novel?

Read more of this story at Slashdot.

What Will Technology Do in 2023?

Looking back at 2022’s technology, the lead technology writer for the New York Times criticized Meta’s $1,500 VR headset and the iPhone’s “mostly unnoticeable improvements.”

But then he also predicted which new tech could affect you in 2023. Some highlights:

– It’s very likely that next year you could have a chatbot that acts as a research assistant. Imagine that you are writing a research paper and want to add some historical facts about World War II. You could share a 100-page document with the bot and ask it to sum up the highlights related to a certain aspect of the war. The bot will then read the document and generate a summary for you….

That doesn’t mean that we’ll see a flood of stand-alone A.I. apps in 2023. It may be more the case that many tools we already use for work will begin building automatic language generation into their apps. Rowan Curran, a technology analyst at the research firm Forrester, said apps like Microsoft Word and Google Sheets could soon embed A.I. tools to streamline people’s work flows.

– In 2023, the V.R. drumbeat will go on. Apple, which has publicly said it will never use the word “metaverse,” is widely expected to release its first headset. Though the company has yet to share details about the product, Apple’s chief executive, Tim Cook, has laid out clues, expressing his excitement about using augmented reality to take advantage of digital data in the physical world. “You’ll wonder how you lived your life without augmented reality, just like today you wonder: How did people like me grow up without the internet?” Mr. Cook said in September to students in Naples.

He added, however, that the technology was not something that would become profound overnight. Wireless headsets remain bulky and used indoors, which means that the first iteration of Apple’s headgear will, similar to many others that preceded it, most likely be used for games. In other words, there will continue to be lots of chatter about the metaverse and virtual (augmented, mixed, whatever-you-want-to-call-dorky-looking) goggles in 2023, but it most likely still won’t be the year that these headsets become widely popular, said Carolina Milanesi, a consumer tech analyst for the research firm Creative Strategies. “From a consumer perspective, it’s still very uncertain what you’re spending your thousand bucks on when you’re buying a headset,” she said. “Do I have to do a meeting with V.R.? With or without legs, it’s not a necessity.”

Read more of this story at Slashdot.

AI-Powered Software Delivery Company Predicts ‘The End of Programming’

Matt Welsh is the CEO and co-founder of Fixie.ai, an AI-powered software delivery company founded by a team from Google and Apple. “I believe the conventional idea of ‘writing a program’ is headed for extinction,” he opines in January’s Communications of the ACM, “and indeed, for all but very specialized applications, most software, as we know it, will be replaced by AI systems that are trained rather than programmed.”

His essay is titled “The End of programming,” and predicts a future will “Programming will be obsolete.”
In situations where one needs a “simple” program (after all, not everything should require a model of hundreds of billions of parameters running on a cluster of GPUs), those programs will, themselves, be generated by an AI rather than coded by hand…. with humans relegated to, at best, a supervisory role…. I am not just talking about things like Github’s CoPilot replacing programmers. I am talking about replacing the entire concept of writing programs with training models. In the future, CS students are not going to need to learn such mundane skills as how to add a node to a binary tree or code in C++. That kind of education will be antiquated, like teaching engineering students how to use a slide rule.

The engineers of the future will, in a few keystrokes, fire up an instance of a four-quintillion-parameter model that already encodes the full extent of human knowledge (and then some), ready to be given any task required of the machine. The bulk of the intellectual work of getting the machine to do what one wants will be about coming up with the right examples, the right training data, and the right ways to evaluate the training process. Suitably powerful models capable of generalizing via few-shot learning will require only a few good examples of the task to be performed. Massive, human-curated datasets will no longer be necessary in most cases, and most people “training” an AI model will not be running gradient descent loops in PyTorch, or anything like it. They will be teaching by example, and the machine will do the rest.

In this new computer science — if we even call it computer science at all — the machines will be so powerful and already know how to do so many things that the field will look like less of an engineering endeavor and more of an an educational one; that is, how to best educate the machine, not unlike the science of how to best educate children in school. Unlike (human) children, though, these AI systems will be flying our airplanes, running our power grids, and possibly even governing entire countries. I would argue that the vast majority of Classical CS becomes irrelevant when our focus turns to teaching intelligent machines rather than directly programming them. Programming, in the conventional sense, will in fact be dead….

We are rapidly moving toward a world where the fundamental building blocks of computation are temperamental, mysterious, adaptive agents…. This shift in the underlying definition of computing presents a huge opportunity, and plenty of huge risks. Yet I think it is time to accept that this is a very likely future, and evolve our thinking accordingly, rather than just sit here waiting for the meteor to hit.

“I think the debate right now is primarily around the extent to which these AI models are going to revolutionize the field,” Welsh says in a video interview. “It’s more a question of degree rather than whether it’s going to happen….

“I think we’re going to change from a world in which people are primarily writing programs by hand to a world in which we’re teaching AI models how to do things that we want them to do… It starts to feel more like a field that focus on AI education and maybe even AI psychiatry. In order to solve these problems, you can’t just assume that people are going to be writing the code by hand.”

Read more of this story at Slashdot.

The Shameful Open Secret Behind Southwest’s Failure? Software Shortcomings

Computer programmer Zeynep Tufekci now writes about the impact of technology on society. In an opinion piece for the New York Times, Tufekci writes on “the shameful open secret” that earlier this week led Southwest airlines to suddenly cancel 5,400 flights in less than 48 hours. “The recent meltdown was avoidable, but it would have cost them.”

Long-time Slashdot reader theodp writes that the piece “takes a crack at explaining ‘technical debt’ to the masses.” Tufekci writes:

Computers become increasingly capable and powerful by the year and new hardware is often the most visible cue for technological progress. However, even with the shiniest hardware, the software that plays a critical role inside many systems is too often antiquated, and in some cases decades old. This failing appears to be a key factor in why Southwest Airlines couldn’t return to business as usual the way other airlines did after last week’s major winter storm. More than 15,000 of its flights were canceled starting on Dec. 22, including more than 2,300 canceled this past Thursday — almost a week after the storm had passed.

It’s been an open secret within Southwest for some time, and a shameful one, that the company desperately needed to modernize its scheduling systems. Software shortcomings had contributed to previous, smaller-scale meltdowns, and Southwest unions had repeatedly warned about it. Without more government regulation and oversight, and greater accountability, we may see more fiascos like this one, which most likely stranded hundreds of thousands of Southwest passengers — perhaps more than a million — over Christmas week.

And not just for a single company, as the problem is widespread across many industries.

“The reason we made it through Y2K intact is that we didn’t ignore the problem,” the piece argues. But in comparison, it points out, Southwest had already experienced another cancellation crisis in October of 2021 (while the president of the pilots’ union “pointed out that the antiquated crew-scheduling technology was leading to cascading disruptions.”) “In March, in its open letter to the company, the union even placed updating the creaking scheduling technology above its demands for increased pay.”

Speaking about this week’s outage, a Southwest spokesman concedes that “We had available crews and aircraft, but our technology struggled to align our resources due to the magnitude and scale of the disruptions.”

But Tufekci concludes that “Ultimately, the problem is that we haven’t built a regulatory environment where companies have incentives to address technical debt, rather than passing the burden on to customers, employees or the next management…. For airlines, it might mean holding them responsible for the problems their miserly approach causes to the flying public.”

Read more of this story at Slashdot.

‘Lifetime Value’ Is Silicon Valley’s Next Buzzword

So long, “total addressable market.” Farewell, “flywheel effect.” Silicon Valley has a new buzzword. As the cost of signing up new customers rises, “lifetime value” is set to become must-use jargon for technology executives, investors and analysts in 2023. Reuters reports: Companies like Uber, DoorDash and Spotify want shareholders to know they can squeeze more revenue out of users than it costs to recruit them. As with previously popular jargon, though, the idea can quickly get garbled. The concept of lifetime value is not new, but a common definition remains elusive. The venture capitalist Bill Gurley defines it as “the net present value of the profit stream of a customer.” Hollywood uses it to estimate the cumulative income from streaming movie titles, after deducting the cost of making the film.

It’s catching on in the tech world. Uber boss Dara Khosrowshahi and his team invoked (PDF) the term seven times during the ride-hailing firm’s investor day. At a similar event in June executives from music streaming service Spotify mentioned (PDF) it 14 times, with another 47 references to the abbreviation LTV. Earnings transcripts for 4,800 U.S.-listed companies analyzed by Bedrock AI show executives and analysts mentioned “lifetime value” over 500 times between October and mid-December, up from just 47 times in three months to March 2019.

The problem is that everyone seems to have a different definition of lifetime value. Food delivery firm DoorDash looks at it as a metric to measure “customer retention, order frequency, and gross profit per order” over a fixed payback period. Uber and its Southeast Asian peer Grab treat it as the ability to bring in one customer and then cross-sell different services at a lower cost. The $49 billion e-commerce firm Shopify defines lifetime value as the total amount of money a customer is expected to spend with the business over the course of an “average business relationship.” But lifetime value isn’t a silver bullet, as Gurley noted a decade ago. As capital becomes more scarce, generating free cash flow remains the most important target. As with previous buzzwords, investors may find that references to lifetime value do more to confound than clarify.

Read more of this story at Slashdot.

Code-Generating AI Can Introduce Security Vulnerabilities, Study Finds

An anonymous reader quotes a report from TechCrunch: A recent study finds that software engineers who use code-generating AI systems are more likely to cause security vulnerabilities in the apps they develop. The paper, co-authored by a team of researchers affiliated with Stanford, highlights the potential pitfalls of code-generating systems as vendors like GitHub start marketing them in earnest. The Stanford study looked specifically at Codex, the AI code-generating system developed by San Francisco-based research lab OpenAI. (Codex powers Copilot.) The researchers recruited 47 developers — ranging from undergraduate students to industry professionals with decades of programming experience — to use Codex to complete security-related problems across programming languages including Python, JavaScript and C.

Codex was trained on billions of lines of public code to suggest additional lines of code and functions given the context of existing code. The system surfaces a programming approach or solution in response to a description of what a developer wants to accomplish (e.g. “Say hello world”), drawing on both its knowledge base and the current context. According to the researchers, the study participants who had access to Codex were more likely to write incorrect and “insecure” (in the cybersecurity sense) solutions to programming problems compared to a control group. Even more concerningly, they were more likely to say that their insecure answers were secure compared to the people in the control.

Megha Srivastava, a postgraduate student at Stanford and the second co-author on the study, stressed that the findings aren’t a complete condemnation of Codex and other code-generating systems. The study participants didn’t have security expertise that might’ve enabled them to better spot code vulnerabilities, for one. That aside, Srivastava believes that code-generating systems are reliably helpful for tasks that aren’t high risk, like exploratory research code, and could with fine-tuning improve in their coding suggestions. “Companies that develop their own [systems], perhaps further trained on their in-house source code, may be better off as the model may be encouraged to generate outputs more in-line with their coding and security practices,” Srivastava said. The co-authors suggest vendors use a mechanism to “refine” users’ prompts to be more secure — “akin to a supervisor looking over and revising rough drafts of code,” reports TechCrunch. “They also suggest that developers of cryptography libraries ensure their default settings are secure, as code-generating systems tend to stick to default values that aren’t always free of exploits.”

Read more of this story at Slashdot.