×
Programming

Is C++ Gaining in Popularity? (i-programmer.info) 106

An anonymous reader shares this report from Dice.com: C++ is enjoying a surge in popularity, according to the latest update to the TIOBE Index, which tracks programming languages' "buzz."

C++ currently sits right behind C and Python on TIOBE's list. "A few months ago, the programming C++ language claimed position 3 of the TIOBE index (at the expense of Java). But C++ has not finished its rise. C seems to be its next victim," added the note accompanying the data... ["At the moment, the gap between the two is only 0.76%."]

Matlab, Scratch and Rust also match their all time high records at respectively positions #10, #12 and #17.

So here, according to TIOBE, are the 10 most popular programmings languages:

1. Python
2. C
3. C++
4. Java
5. C#
6. JavaScript
7. Visual Basic
8. SQL
9. PHP
10. MATLAB

The site I Programmer digs deeper: C++ was the only one of the top four languages to see a positive year-on-year change in its percentage rating — adding 0.79% to stand at 10.8%. Python had the smallest loss of the entire Top 20, -0.01% leaving it with a share of 13,42% while Visual Basic had the greatest loss at -2.07%. This, combined with JavaScript gaining 1.34%, led to JavaScript overtaking it to occupy #6, its highest ever ranking in the TIOBE Index.
They also note that COBOL "had a 3-month rise going from a share of 0.41% in April to 0.86% in July which moved it into #20 on the index."
Supercomputing

Cerebras To Enable 'Condor Galaxy' Network of AI Supercomputers 20

Cerebras Systems and G42 have introduced the Condor Galaxy project, a network of nine interconnected supercomputers designed for AI model training with a combined performance of 36 FP16 ExaFLOPs. The first supercomputer, CG-1, located in California, offers 4 ExaFLOPs of FP16 performance and 54 million cores, focusing on Large Language Models and Generative AI without the need for complex distributed programming languages. AnandTech reports: CG-2 and CG-3 will be located in the U.S. and will follow in 2024. The remaining systems will be located across the globe and the total cost of the project will be over $900 million. The CG-1 supercomputer, situated in Santa Clara, California, combines 64 Cerebras CS-2 systems into a single user-friendly AI supercomputer, capable of providing 4 ExaFLOPs of dense, systolic FP16 compute for AI training. Based around Cerebras's 2.6 trillion transistor second-generation wafer scale engine processors, the machine is designed specifically for Large Language Models and Generative AI. It supports up to 600 billion parameter models, with configurations that can be expanded to support up to 100 trillion parameter models. Its 54 million AI-optimized compute cores and massivefabric network bandwidth of 388 Tb/s allow for nearly linear performance scaling from 1 to 64 CS-2 systems, according to Cerebras. The CG-1 supercomputer also offers inherent support for long sequence length training (up to 50,000 tokens) and does not require any complex distributed programming languages, which is common in case of GPU clusters.

This supercomputer is provided as a cloud service by Cerebras and G42 and since it is located in the U.S., Cerebras and G42 assert that it will not be used by hostile states. CG-1 is the first of three 4 FP16 ExaFLOP AI supercomputers (CG-1, CG-2, and CG-3) created by Cerebras and G42 in collaboration and located in the U.S. Once connected, these three AI supercomputers will form a 12 FP16 ExaFLOP, 162 million core distributed AI supercomputer, though it remains to be seen how efficient this network will be. In 2024, G42 and Cerebras plan to launch six additional Condor Galaxy supercomputers across the world, which will increase the total compute power to 36 FP16 ExaFLOPs delivered by 576 CS-2 systems. The Condor Galaxy project aims to democratize AI by offering sophisticated AI compute technology in the cloud.
"Delivering 4 exaFLOPs of AI compute at FP16, CG-1 dramatically reduces AI training timelines while eliminating the pain of distributed compute," said Andrew Feldman, CEO of Cerebras Systems. "Many cloud companies have announced massive GPU clusters that cost billions of dollars to build, but that are extremely difficult to use. Distributing a single model over thousands of tiny GPUs takes months of time from dozens of people with rare expertise. CG-1 eliminates this challenge. Setting up a generative AI model takes minutes, not months and can be done by a single person. CG-1 is the first of three 4 ExaFLOP AI supercomputers to be deployed across the U.S. Over the next year, together with G42, we plan to expand this deployment and stand up a staggering 36 exaFLOPs of efficient, purpose-built AI compute."
Programming

Most Outsourced Coders In India Will Be Gone In 2 Years Due To AI, Stability AI Boss Predicts (cnbc.com) 85

Most outsourced programmers in India will see their jobs wiped out in the next year or two, Stability AI CEO Emad Mostaque said. CNBC reports: Mostaque, on a call with UBS analysts, said that most of the country's outsourced coders will lose their jobs as the effects of AI mean that it is now possible for software to be developed with far fewer people. "I think that it affects different types of jobs in different ways," Mostaque said on a call with analysts at the Swiss investment bank last week. "If you're doing a job in front of a computer, and no one ever sees you, then it's massively impactful, because these models are like really talented grads."

According to Mostaque, not everyone will be affected in the same way, however. That is due in no small part to differing rules and regulations around the world. Countries with stronger labor laws, like France, will be less likely to see such an impact, for example. In India, Mostaque said, "outsourced coders up to level three programmers will be gone in the next year or two, whereas in France, you'll never fire a developer." "So it affects different models in different countries in different ways in different sectors."

Mostaque reiterated a previous statement he made saying that there will be "no more programmers" in five years' time -- however, he caveated this to say that he meant coders in the traditional sense. "Why would you have to write code where the computer can write code better? When you deconstruct the programming thing from bug testing to unit testing to ideation, an AI can do that, just better," Mostaque said. "But it won't be doing it automatically, it will be AI 'co-pilots,'" Mostaque said. "That means less people are needed for classical programming, but then are they needed for other things? This is the question and this is the balance that we have to understand, because different areas are also affected differently."

Music

Plex's Winamp-inspired Music Player Plexamp is Now Free (techcrunch.com) 46

Plexamp, the music player originally incubated by the Labs division of media company Plex, is now free, the company announced today. From a report: The project was first launched in 2017 as Plex's own spin on the classic Winamp media player app, offering visualizations to accompany your tunes, tools for programming mixes, and more recently, a ChatGPT-powered "Sonic Sage" feature that builds unique playlists from users' music libraries. However, after its expansion from desktop to mobile, Plexamp was only available to subscribers.

Now, Plex says the Plexamp app will become free, allowing users to play tracks from their own library or the TIDAL music streaming service with high-quality audio and support for lossless audio. The app also includes gapless playback, loudness leveling, and smooth transitions between tracks, among other things. In addition to Library Radio, a feature used to rediscover your music, users can create playlists with Plexamp to match their current mood: like "brooding, cathartic, confident, intense, playful, poignant, swaggering, and wistful," the company says.

AI

Anthropic Releases a New Version of Its ChatGPT Rival, Claude (bloomberg.com) 23

Anthropic, an artificial intelligence startup positioning itself as the builder of a safer kind of chatbot, has released a new version of its AI bot, named Claude. From a report: Anthropic said that Claude 2 is available to anyone in the US or UK online at claude.ai, and businesses can access it via an application programming interface. The new release on Tuesday comes several months after Anthropic began offering an earlier version of Claude to businesses that wanted to add it to their products. Previously, the bot was tested by a handful of companies including Quora, which built it into an app called Poe that lets users ask questions.

Like its predecessor, Claude 2 is built atop a large language model and can be used for written tasks like summarizing, searching, answering questions and coding. Both models can currently take in large chunks of text -- a user can ask it to summarize a book, for instance -- though Claude 2 can generate longer responses than its predecessor. Responses can reach up to about 3,000 words, according to data provided by the company. Claude 2 will also offer more accurate responses on some topics, such as coding and grade-school-level math, the company said. Anthropic's goal has been for Claude to be less susceptible than other chatbots to manipulation.

Anime

Hulu Launches Adult Animation, Anime Hub Animayhem (variety.com) 24

According to Variety, Hulu is launching a new sub-brand focused on adult animation and anime called Animayhem. From the report: The new hub is meant to capitalize on Hulu's already popular lineup of adult animation and anime shows. Series like "American Dad," "Bob's Burgers," "Family Guy," "Futurama" and "King of the Hill" consistently rank among the service's top 10 shows based on hours streamed, per Hulu. So far this year, over one billion hours of adult animation content has been streamed on Hulu, along with over 288 million hours of anime content, the streamer claims.

Hulu currently has 46 adult animated series, which adds up to 174 seasons and 2,600 episodes. For anime, Hulu has 17 films and 272 series, adding up to 435 seasons and 18,400 episodes (including subtitled and English dubbed versions of episodes). As part of the launch, Hulu is debuting a range of ads for Animayhem, which they describe as the "Animation Destination," one of which can be seen below.
"When you have the number one offering in adult animation and anime of any major streaming service, creating this destination is obvious. We know exactly where we can meet these fans, because they're already here," said Barrie Gruner, Hulu's executive vice president of marketing and publicity, in an interview with Variety.

"I would say that this brand really cements Hulu as the ultimate streaming destination for animation and we're not going to achieve that with single title campaigns," Gruner said. "This is truly an intersection with our original programming and our library."

Animayhem will also be coming to San Diego Comic-Con via an immersive experience dubbed "Hulu Animayhem: Into the Second Dimension."
Programming

Why Are There So Many Programming Languages? (acm.org) 160

Long-time Slashdot reader theodp writes: Recalling a past Computer History Museum look at the evolution of programming languages, Doug Meil ponders the age-old question of Why Are There So Many Programming Languages? in a new Communications of the ACM blog post.

"It's worth noting and admiring the audacity of PL/I (1964)," Meil writes, "which was aiming to be that 'one good programming language.' The name says it all: Programming Language 1. There should be no need for 2, 3, or 4. [Meil expands on this thought in Lessons from PL/I: A Most Ambitious Programming Language.] Though PL/I's plans of becoming the Highlander of computer programming didn't play out like the designers intended, they were still pulling on a key thread in software: why so many languages? That question was already being asked as far back as the early 1960's."

One of PL/I's biggest fans was Digital Research Inc. (DRI) founder Gary Kildall, who crafted the PL/I-inspired PL/M (Programming Language for Microcomputers) in 1973 for Intel. But IBM priced PL/I higher than the languages it sought to replace, contributing to PL/I's failure to gain traction. (Along the lines of how IBM's deal with Microsoft gave rise to a price disparity that was the undoing of Kildall's CP/M OS, bundled with every PC in a 'non-royalty' deal. Windows was priced at $40 while CP/M was offered 'a la carte' at $240.) As a comp.lang.pl1 poster explained in 2006, "The truth of the matter is that Gresham's Law: 'Bad money drives out good' or Ruskin's principle: 'The hoi polloi always prefer an inferior, cheap product over a superior, more expensive one' are what govern here."

Perl

Perl 5.38 Released with New Experimental Syntax for Defining Object Classes (phoronix.com) 48

Perl 5.38 was released this week "after being in development for more than one year," reports Phoronix. "Perl 5.38 brings a new experimental syntax for defining object classes where per-instance data is stored in 'field' variables that behave like lexicals."

"Maybe, just maybe, the new features introduced into the language in this newest version will attract much sought new talent," writes the site I Programmer, noting the argument that Perl is installed by default everywhere — and has the "fun factor... The class keyword is part of the plan to bring effective object-oriented programming to the Perl core while still keeping Perl being Perl."

The Perl docs warn that "This remains a new and experimental feature, and is very much still under development. It will be the subject of much further addition, refinement and alteration in future releases." But "Since Perl 5, support for objects revolved around the concept of blessing references with a package name," notes updated documentation, which points out this new class syntax "isn't a bless wrapper, but a completely new system built right into the perl interpreter." The class keyword declares a new package which is intended to be a class... classes automatically get a constructor named new... Just like with other references, when object reference count reaches zero it will automatically be destroyed.
Phoronx notes that Perl 5.38 also brings a new PERL_RAND_SEED environment variable "for controlling seed behavior for random number generation," along with some new APIs. And I Programmer adds that Perl 5.38 also adds support for Unicode 15.0, adding 4, 489 characters, for a total of 149,186 characters. Other additions include enhanced regular expressions, plus defined-or and logical-or assignment default expressions in signatures.
Businesses

Amazon CEO Asks His Hollywood Studio To Explain Its Big Spending (bloomberg.com) 110

Amazon CEO Andy Jassy is taking a hard look at how much the company's Hollywood studio spends on original TV programming. From a report: In recent weeks, he has asked executives for detailed budget analyses of some of their biggest shows, according to people familiar with the matter, scrutinizing the studio's ballooning costs and mixed track record with audiences. The world's largest online retailer is engaged in a companywide cost-cutting program, with plans to eliminate at least 27,000 jobs. Across Amazon, Jassy has also jettisoned 37 different projects deemed unnecessary.

The Hollywood studio, which has spent tens of billions of dollars on original programming over the last decade, is an obvious place to look for savings. Last year, Amazon spent $7 billion on original shows, licensed programs and sports, up from $5 billion the year before. Only Netflix and Disney spend more on streaming. In the past nine months, Amazon has released at least a half-dozen pricey series that failed to deliver huge audiences. Daisy Jones & the Six, The Power, Dead Ringers and The Peripheral all cost more than $100 million to produce but failed to crack Nielsen's list of the 10 most-watched streaming programs in the US. Even The Rings of Power ($400 million-plus), a show that attracted a large audience, failed to hold on to most of its viewers over the course of the season, according to The Hollywood Reporter.

The Almighty Buck

US Might Finally Force Cable-TV Firms To Advertise Their Actual Prices (arstechnica.com) 67

The Federal Communications Commission (FCC) has proposed new rules to crack down on hidden fees charged by cable and satellite video providers. "My administration's top priority is lowering the cost of living for the middle class, and that includes cracking down on companies' use of junk fees to hide true costs from families, who end up paying more as a result," Biden said in a statement on Tuesday. Ars Technica reports: As Biden noted, the FCC "proposed a new rule that would require cable and satellite TV providers to give consumers the all-in price for the service they're offering up front." The proposed rule would force companies like Comcast, Charter Spectrum, and DirecTV to publish more accurate prices. Biden continued: "Too often, these companies hide additional junk fees on customer bills disguised as "broadcast TV" or "regional sports" fees that in reality pay for no additional services. These fees really add up: according to one report, they increase customer bills by nearly 25 percent of the price of base service."

FCC Chairwoman Jessica Rosenworcel first floated pricing transparency rules for the TV services offered by cable and satellite companies in March. That effort took a step forward on Tuesday when the commission approved a Notice of Proposed Rulemaking (NPRM) that seeks public comment on rules that would force video providers to offer accurate prices in advertising. "Consumers who choose a video service based on an advertised monthly price may be surprised by unexpected fees related to the cost of video programming that raise the amount of the bill significantly," the NPRM said. The cable and satellite TV companies' practice of listing "Broadcast TV" and "Regional Sports Network" fees separately from the advertised price "can be potentially misleading and interpreted as a government-imposed tax or fee, instead of a company-imposed service fee increase," and make it hard for customers to compare prices across providers, the FCC said.

The docket is available here, and comments will be accepted for 60 days after the NPRM is published in the Federal Register. The FCC said its proposal "would require cable operators and DBS [direct broadcast satellite] providers to clearly and prominently display the total cost of video programming service." The FCC is also seeking comment on whether it has the authority to impose similar requirements on other types of video providers. But Rosenworcel reportedly said in a congressional hearing that the FCC's authority under US law doesn't extend to streaming services.

Intel

Intel's New Font For Low-Vision Developers Is Causing Design Drama For Coders (fastcompany.com) 96

Elissaveta M. Brandon writes via Fast Company: There's a new font in town -- and it's already causing rifts on Reddit. The font is called Intel One Mono, and as its name implies, it was designed by tech giant Intel, together with New York-based type design practice Frere-Jones Type and marketing agency VMLY&R. It joins a group of monospaced fonts designed primarily for developers -- think JetBrains Mono, Fira Code, and Consolas. By definition, monospaced fonts consist of characters that have the same width and occupy the same horizontal space, making it easy for coders and programmers to tell the difference between long strings of characters. But here's where Intel One Mono stands out: it was designed with and for low-vision developers. (It's free to download on GitHub and will soon be available on Google Fonts, too.)

To ensure the font was legible and readable to its target audience, the team ran more than a dozen "live testing sessions" with visually impaired developers who were asked to write code using Intel One Mono. [...] Some of the feedback the designers received was particularly surprising. For example, some people were struggling to tell apart a capital "M" from a capital "N," most likely because both letters have two vertical stems and some diagonals in between, which can be confusing. To make the letters more legible, the designers sloped the vertical stems on the "M" so it looks close to an inverted W. "The point at which the two diagonals meet in the middle gets shifted up to make it clearly a V shape in the middle, and then the two verticals get flared out a little bit to give it slightly more differentiable shape from the capital N," says Fred Shallcrass, a type designer at Frere-Jones Type.

Similar challenges kept coming back with the "x" and the "y" which people struggled to distinguish, and the "e" and the "c." In every instance, the designers meticulously tweaked the letters to make them highly distinctive, resulting in a fairly idiosyncratic font where every glyph is as different as possible from the other -- all the way down to the curly brackets, which can best be described as extra curly. This brings us to that Reddit rift. "This font would be great were it not for those curly braces," one person wrote. "For someone that hates fonts sometimes because of curly brackets not being clear and evident, I'm officially switching to this font set because of the curly brackets," wrote another. The developers were equally torn, but the designers stand by them.
"Part of our thinking in negotiating those responses is that reinforcing the identity of any shape is not just amplifying what is unique about that letter, but also making it clearly not some other letter, so foreclosing any confusion," says Tobias Frere-Jones, the founder and lead designer at his eponymous studio. "If there's a thing the curly braces do, which is that extra back and forth movement, the parentheses don't do that, the brackets don't do that, therefore these ought to do a lot of that."
Bitcoin

Mastercard Submits Fresh Trademark Application For Crypto Tech (crypto.news) 18

According to a recently discovered patent application, Mastercard plans to develop software optimized for bitcoin and blockchain transactions. The second-largest payment-processing corporation also aims to facilitate crypto-based transactions by reducing connections between virtual asset service providers. Crypto News reports: The trademark application is a fascinating window into Mastercard's plans for the future of digital currency. Details have been revealed about creating a downloadable application programming interface (API) designed to verify transactions inside blockchain networks and ease the handling or trading of cryptocurrency. By standardizing this API software, communication between VASPs may be streamlined and crypto transactions easier. Mastercard wants to set up a platform for financial institutions to exchange customer information to verify compliance. This new step is significant for Mastercard's fast-growing presence in the cryptocurrency sector. The corporation announced its intention to offer a limited number of cryptocurrencies on its network in February 2021.
AI

Hey Alexa, What Should Students Learn About AI? (nytimes.com) 22

Long-time Slashdot reader theodp writes: While schools debate what to teach students about powerful new A.I. tools, tech giants, universities and nonprofits are intervening with free lessons," writes the NY Times reports in Hey, Alexa, What Should Students Learn About AI?
Senior Amazon executive Rohit Prasad visited a school in Boston called STEM Academy to observe an Amazon-sponsored AI lesson using Alexa, according to the article, "And he assured the Dearborn students there would soon be millions of new jobs in A.I." "We need to create the talent for the next generation," Mr. Prasad, the head scientist for Alexa, told the class. "So we are educating about A.I. at the earliest, grass-roots level."

A few miles away, Sally Kornbluth, the president of the Massachusetts Institute of Technology, was delivering a more sobering message about A.I. to students from local schools who had gathered at Boston's Kennedy Library complex for a workshop on A.I. risks and regulation. "Because A.I. is such a powerful new technology, in order for it to work well in society, it really needs some rules," Dr. Kornbluth said. "We have to make sure that what it doesn't do is cause harm."

The same-day events — one encouraging work in artificial intelligence and the other cautioning against deploying the technology too hastily — mirrored the larger debate currently raging in the United States over the promise and potential peril of A.I. Both student workshops were organized by an M.I.T. initiative on "responsible A.I." whose donors include Amazon, Google and Microsoft.

The article emphasizes that schools face a big question: Should they teach AI programming and other AI-related skills employers will seek? "Or should students learn to anticipate and mitigate A.I. harms?"

Last week, Amazon agreed to pay $25 million to settle federal charges that it had indefinitely kept children's voice recordings, violating the federal online children's privacy law. The company said it disputed the charges and denied that it had violated the law. The company noted that customers could review and delete their Alexa voice recordings. But the one-hour Amazon-led workshop did not touch on the company's data practices.
Programming

Is AI an Excuse for Not Learning To Code? (acm.org) 133

Long-time Slashdot reader theodp writes: Y Combinator founder Paul Graham last week took to Twitter to lament those who use AI or other excuses for not learning to code. "A generation ago some people were saying there was no point in learning to program because all the programming jobs would be outsourced to India," Graham wrote. "Now they're saying you don't need to because AI will do it all. If you don't want to learn to program, you can always find a reason."

BloomTech Coding Bootcamp CEO Austen Allred this week doubled-down on Graham's tweet, offering his own history of excuses people have made for not learning to code... Allred's tweet reads:

"Don't learn to code. Soon GUIs will do it all for you." — 1985

"Don't learn to code. Soon that will all be done offshore for pennies." — 2003

"Don't learn to code. Soon nocode tools will do it all for you." — 2015

"Don't learn to code. Soon AI will do it all for you." — 2023

Among the many retweeting Allred's cautionary message was Code.org, the tech-backed nonprofit that aims to make computer science a high school graduation requirement by 2030, whose CEO also replied to Graham with a reassuring tweet suggesting people's days of being able to avoid learning to code will soon be over. "Now that 27 states require that every school must teach computer science, and 7 states require a CS course to graduate high school," explained Code.org CEO Hadi Partovi, "the argument is basically behind us. Computer science won."

On a related note, this month in Communications of the ACM, a CS professor shared their own contrary opinion about the possibility of a professional programmer using AI assistants to do a better job.

"It doesn't work." I would love to have an assistant who keeps me in check, alerting me to pitfalls and correcting me when I err. A effective pair-programmer. But that is not what I get. Instead, I have the equivalent of a cocky graduate student, smart and widely read, also polite and quick to apologize, but thoroughly, invariably, sloppy and unreliable. I have little use for such supposed help...

Fascinating as they are, AI assistants are not works of logic; they are works of words. Large language models: smooth talkers (like the ones who got all the dates in high school). They have become incredibly good at producing text that looks right. For many applications that is enough. Not for programming.

Programming

92% of Programmers Are Using AI Tools, Says GitHub Developer Survey 67

An anonymous reader quotes a report from ZDNet: [A]ccording to a new GitHub programmer survey, "92% of US-based developers are already using AI coding tools both in and outside of work." GitHub partnered with Wakefield Research to survey 500 US-based enterprise developers. They found that 70% of programmers believe AI is providing significant benefits to their code. Specifically, developers said AI coding tools can help them meet existing performance standards with improved code quality, faster outputs, and fewer production-level incidents.

This is more than just people working on external open-source projects or just fooling around. Only 6% of developers said they solely use these tools outside of work. In other words, today, AI programming tools are part and parcel of modern business IT. Why has this happened so quickly? It's all about the programmers' bottom line. Developers say AI coding tools help them meet existing performance standards with improved code quality, faster outputs, and fewer production-level incidents. It's also all about simply producing more lines of code.
"Engineering leaders will need to ask whether measuring code volume is still the best way to measure productivity and output," added Inbal Shani, GitHub's chief product officer. "Ultimately, the way to innovate at scale is to empower developers by improving their productivity, increasing their satisfaction, and enabling them to do their best work -- every day."

According to the survey, "Developers want to upskill, design solutions, get feedback from end users, and be evaluated on their communication skills."

"In other words, generating code with AI is a means to an end, not an end to itself," writes ZDNet's Steven Vaughan-Nichols. "Developers believe they should be judged on how they handle those bugs and issues, which is more important to performance than just lines of code. [...] Yes, you can have ChatGPT write a program for you, but if you don't understand what you're doing in the first place or the code you're 'writing,' the code will still be garbage. So, don't think for a minute that just because you can use ChatGPT to write a Rust bubble-sort routine, it means you're a programmer now, You're not."
Programming

Google Home's Script Editor Is Now Live (theverge.com) 23

Google has launched its script editor tool, offering advanced automations for Google Home-powered smart homes. The Verge reports: Available starting Tuesday, June 13th, to those in the Google Home public preview, the script editor is part of Google's new home.google.com web interface, which also has live feeds for any Nest cams on your account. The script editor will be coming to the new Google Home app preview starting June 14th. There's no date for general availability.

Along with allowing for multiple starters and actions, the script editor adds more advanced conditions. For example, you can set an automation to run only if the TV is on and it's after 6PM but before midnight. The script editor automations are created in the new Google Home web interface, you can apply for the public preview here.

The script editor allows you to do everything you can in the Home app when setting up automations, plus "more than 100 new features and capabilities to fit your unique understanding of your home and what you want it to do," according to a blog post by Anish Kattukaran, director of product management at Google Home. This includes access to nearly 100 starters and actions, including Matter sensors -- something not currently possible in the Home app. For example, an Eve Motion sensor connected via Matter to Google Home can't currently be used as a starter for automations in the Home app but can be used as one in the script editor.
Google has some example automations that you can view here.
Programming

Does the New 'Mojo' Programming Language Offer a Faster Superset of Python? (infoworld.com) 71

InfoWorld explores how the new Mojo program language "resembles Python, how it's different, and what it has to offer." The newly unveiled Mojo language is being promoted as the best of multiple worlds: the ease of use and clear syntax of Python, with the speed and memory safety of Rust. Those are bold claims, and since Mojo is still in the very early stages of development, it will be some time before users can see for themselves how the language lives up to them. But Mojo's originator — a company named Modular — has provided early access [through a limited-enrollment preview program] to an online playground: a Jupyter Notebook environment where users can run Mojo code and learn about the language's features and behavior...

Mojo can be described as a "superset" of Python. Programs written in Python are valid Mojo programs, although some Python behaviors haven't yet been implemented... It's also possible to use the actual Python runtime for working with existing Python modules, although there is a performance cost. When Mojo introduces new syntax, it's for system-level programming features, chiefly manual memory handling. In other words, you can write Python code (or something almost exactly like it) for casual use cases, then use Mojo for more advanced, performance-intensive programming scenarios... Mojo's other big difference from Python is that Mojo's not interpreted through a runtime, as Python is. Mojo is compiled ahead-of-time to machine-native code, using the LLVM toolchain. To that end, the best performance comes from using features specific to Mojo. Python features are likely to come at the cost of emulating Python's dynamic behaviors, which are inherently slow — or again, by just using the Python runtime.

Many of Mojo's native language features do one of two things. They're either entirely new features not found in Python at all, or expansions of a Python feature that make it more performant, although with less of Python's dynamism.

For example, Mojo has its own fn keyword which defines a function with explicitly-typed and immutable-by-default arguments, and its own struct keyword which is less like a Python class and more like its C/C++ and Rust counterpart "with fixed layouts determined at compile time but optimized for machine-native speed."

But "At a glance, the code closely resembles Python. Even the new Mojo-specific keywords integrate well with existing Python syntax, so you can run your eye down the code and get a general idea of what's happening." And then there's the speed... The notebook demos also give examples of how Mojo code can be accelerated via parallelism, vectorizing, and "tiling" (increasing cache locality for operations). One of the demos, a 128x128 matrix multiplication demo, yielded a claimed 17-times speedup over Python (using the Python runtime in the Mojo playground) by simply running as-is with no special modification. Mojo added 1866x speedup by adding type annotations, 8500x speedup by adding vectorized operations, and 15000x speedup by adding parallelization.
AI

Will Productivity Gains from AI-Generated Code Be Offset by the Need to Maintain and Review It? (zdnet.com) 95

ZDNet asks the million-dollar question. "Despite the potential for vast productivity gains from generative AI tools such as ChatGPT or GitHub Copilot, will technology professionals' jobs actually grow more complicated? " People can now pump out code on demand in an abundance of languages, from Java to Python, along with helpful recommendations. Already, 95% of developers in a recent survey from Sourcegraph report they use Copilot, ChatGPT, and other gen AI tools this way.

But auto-generating new code only addresses part of the problem in enterprises that already maintain unwieldy codebases, and require high levels of cohesion, accountability, and security.

For starters, security and quality assurance tasks associated with software jobs aren't going to go away anytime soon. "For programmers and software engineers, ChatGPT and other large language models help create code in almost any language," says Andy Thurai, analyst with Constellation Research, before talking about security concerns. "However, most of the code that is generated is security-vulnerable and might not pass enterprise-grade code. So, while AI can help accelerate coding, care should be taken to analyze the code, find vulnerabilities, and fix it, which would take away some of the productivity increase that AI vendors tout about."

Then there's code sprawl. An analogy to the rollout of generative AI in coding is the introduction of cloud computing, which seemed to simplify application acquisition when first rolled out, and now means a tangle of services to be managed. The relative ease of generating code via AI will contribute to an ever-expanding codebase — what the Sourcegraph survey authors refer to as "Big Code". A majority of the 500 developers in the survey are concerned about managing all this new code, along with code sprawl, and its contribution to technical debt. Even before generative AI, close to eight in 10 say their codebase grew five times over the last three years, and a similar number struggle with understanding existing code generated by others.

So, the productivity prospects for generative AI in programming are a mixed bag.

AI

Is Self-Healing Code the Future of Software Development? (stackoverflow.blog) 99

We already have automated processes that detect bugs, test solutions, and generate documentation, notes a new post on Stack Overflow's blog. But beyond that, several developers "have written in the past on the idea of self-healing code. Head over to Stack Overflow's CI/CD Collective and you'll find numerous examples of technologists putting this ideas into practice."

Their blog post argues that self-healing code "is the future of software development." When code fails, it often gives an error message. If your software is any good, that error message will say exactly what was wrong and point you in the direction of a fix. Previous self-healing code programs are clever automations that reduce errors, allow for graceful fallbacks, and manage alerts. Maybe you want to add a little disk space or delete some files when you get a warning that utilization is at 90% percent. Or hey, have you tried turning it off and then back on again?

Developers love automating solutions to their problems, and with the rise of generative AI, this concept is likely to be applied to both the creation, maintenance, and the improvement of code at an entirely new level... "People have talked about technical debt for a long time, and now we have a brand new credit card here that is going to allow us to accumulate technical debt in ways we were never able to do before," said Armando Solar-Lezama, a professor at the Massachusetts Institute of Technology's Computer Science & Artificial Intelligence Laboratory, in an interview with the Wall Street Journal. "I think there is a risk of accumulating lots of very shoddy code written by a machine," he said, adding that companies will have to rethink methodologies around how they can work in tandem with the new tools' capabilities to avoid that.

Despite the occasional "hallucination" of non-existent information, Stack Overflow's blog acknowledges that large-language models improve when asked to review their response, identify errors, or show its work.

And they point out the project manager in charge of generative models at Google "believes that some of the work of checking the code over for accuracy, security, and speed will eventually fall to AI." Google is already using this technology to help speed up the process of resolving code review comments. The authors of a recent paper on this approach write that, "As of today, code-change authors at Google address a substantial amount of reviewer comments by applying an ML-suggested edit. We expect that to reduce time spent on code reviews by hundreds of thousands of hours annually at Google scale. Unsolicited, very positive feedback highlights that the impact of ML-suggested code edits increases Googlers' productivity and allows them to focus on more creative and complex tasks...."

Recently, we've seen some intriguing experiments that apply this review capability to code you're trying to deploy. Say a code push triggers an alert on a build failure in your CI pipeline. A plugin triggers a GitHub action that automatically send the code to a sandbox where an AI can review the code and the error, then commit a fix. That new code is run through the pipeline again, and if it passes the test, is moved to deploy... Right now his work happens in the CI/CD pipeline, but [Calvin Hoenes, the plugin's creator] dreams of a world where these kind of agents can help fix errors that arise from code that's already live in the world. "What's very fascinating is when you actually have in production code running and producing an error, could it heal itself on the fly?" asks Hoenes...

For now, says Hoenes, we need humans in the loop. Will there come a time when computer programs are expected to autonomously heal themselves as they are crafted and grown? "I mean, if you have great test coverage, right, if you have a hundred percent test coverage, you have a very clean, clean codebase, I can see that happening. For the medium, foreseeable future, we probably better off with the humans in the loop."

Last month Stack Overflow themselves tried an AI experiment that helped users to craft a good title for their question.
Intel

Intel Open Sources New 'One Mono' Font for Programmers (github.com) 51

Intel has announced Intel One Mono, a new font catering to "the needs of developers" with an "expressive" monospace for clarity and legibility" It's easier to read, and available for free, with an open-source font license.

Identifying the typographically underserved low-vision developer audience, Frere-Jones Type designed the Intel One Mono typeface in partnership with the Intel Brand Team and VMLY&R, for maximum legibility to address developers' fatigue and eyestrain and reduce coding errors. A panel of low-vision and legally blind developers provided feedback at each stage of design.

The Linux blog OMG! Ubuntu calls the new font "pretty decent," adding that "Between IBM Plex Mono, Hack, Fira Code, and JetBrains Mono I think we Linux users are spoilt for choice when it comes to open-source monospace fonts that look good and work great.

"Still, there's always room for more, right...?" Better yet, it's not only free to download and use but free to edit, and free to redistribute... Overall, I think Intel One Mono looks great, especially in a text editor (GUI or CLI). There's a noticeable upper and lower margin to the font that in dense text situations allows text to breathe, but in some terminal tools, like Neofetch, the gaps can seem a bit too happy.
The Intel One Mono repository on GitHub includes instructions for activating the font in VSCode and Sublime Text, and lists some extra features accessible in some applications and via CSS:
  • There is an option for a raised colon, either applied contextually between numbers or activated generally.
  • Superior/superscript and inferior/subscript figures are included via their Unicode codepoints, or you can produce them from the default figures via the sups (Superscript), subs (Subscript), and si (Scientific Inferior) features.
  • Fraction numerals are similarly available via the numr (Numerator) and dnom (Denominator) features. A set of premade fractions is also available in the fonts.

Slashdot Top Deals