Tesla CEO Elon Musk showcases humanoid robot at event
Musk says Optimus will be an “extremely capable robot,” unlike other humanoid robots that don’t have the intelligence to navigate the world by themselves
Updated 01 October 2022
SAN FRANCISCO: Tesla CEO Elon Musk showcased his much-touted humanoid robot ‘Optimus’ at the electric vehicle maker’s “AI Day” event on Friday.
The billionaire has said a robot business will be worth more than its cars, hoping to expand beyond self-driving cars that have not yet become a reality despite his repeated promises.
A prototype of the robot walked on stage and waved to the seated audience. A video of the robot carrying a box, watering plants and moving metal bars in the automaker’s factory was shown.
“Our goal is to make a useful humanoid robot as quickly as possible,” Musk said at the event being held at a Tesla office in Palo Alto, California.
“There’s still a lot of work to be done to refine Optimus and prove it.”
Musk said currently humanoid robots are “missing a brain,” saying they don’t have the intelligence to navigate the world by themselves, and they are also very expensive and made in low volume.
By contrast, he said, the Optimus will be an “extremely capable robot,” to be made in very high volume, probably, ultimately millions of units and is expected to cost much less than a car, at under $20,000.
Musk is also expected to discuss Tesla’s long-delayed self-driving technology. In May, Musk said that the world’s most valuable car maker would be “worth basically zero” without achieving full self-driving capability, and it faces growing regulatory probes, as well as technological hurdles.
“There will be lots of technical detail & cool hardware demos,” Musk wrote on Twitter late on Wednesday, adding the event was aimed at recruiting engineers.
Tesla’s live demonstration record is mixed. Launches typically draw cheers, but in 2019 when Musk had an employee hurl a steel ball at the armored window of a new electric pickup truck, the glass cracked.
The key test for the robot is whether it can handle unexpected situations.
Musk announced Tesla’s plan for humanoid robots at its AI day in August last year and delayed this year’s event from August to have its robot prototype working, with a plan to start production possibly next year.
Tesla teased the unveiling of the bot on social media with an image of metallic robotic hands making a heart shape. But building human-like, versatile hands that can manipulate different objects is extremely challenging, said Heni Ben Amor, a robotics professor at Arizona State University.
Initially, Optimus, an allusion to the powerful and benevolent leader of the Autobots in the Transformers media franchise, would perform boring or dangerous jobs, including moving parts around Tesla factories or attaching a bolt to a car with a wrench, according to Musk.
“There’s so much about what people can do dexterously that’s very, very hard for robots. And that’s not going to change whether the robot is a robot arm or whether it’s in the shape of a humanoid,” Jonathan Hurst, chief technology officer at Agility Robotics, a humanoid robot firm, told Reuters.
Musk has said that in the future robots could be used in homes, making dinners, mowing the lawn and caring for the elderly, and even becoming a “buddy” for humans or a sex partner.
He is due at Friday’s event to give updates on Tesla’s much-delayed plan to launch self-driving cars, and on its high-speed computer, Dojo, which was unveiled last year and the company has said is integral to its development of self-driving technology.
Musk has said he expects Tesla will achieve full self-driving this year and mass produce a robotaxi with no steering wheel or pedal by 2024.
At an “Autonomy” event in 2019, Musk promised 1 million robotaxis by 2020 but has yet to deliver such a car.
KARACHI: Pakistani developers moved up to 16th place in 2023 from 27th in 2018 in terms of app downloads, with apps developed in the South Asian country downloaded four billion times in 2022, Google said on Thursday.
Pakistan’s IT exports during the first half of the current fiscal year (FY23) increased by two percent to $1.3 billion, contributing 38% to the overall services’ export, according to research by Arif Habib, a leading Pakistani securities brokerage, investment banking and research firm.
In 2022, 65 apps from 24 Pakistani developers reached the top 10 rankings in the gaming and apps verticals on both Play Store and App Store, according to Google.
“In 2022, apps developed in Pakistan were downloaded a staggering 4 billion times worldwide, marking a compound annual growth rate (CAGR) of 30% from 2018 to 2020,” the company said.
Google said it recently held the “first and the biggest” offline apps and games-related event in Pakistan’s eastern city of Lahore, bringing together more than 500 developers and industry leaders to learn about scaling up Pakistani apps and games globally.
“To help Pakistani studios and developers further unlock digital opportunities, Google is expanding its support for the apps industry in Pakistan through [events like] Think Apps, which are designed to bring in-depth knowledge and trends from global experts to local developers,” the Google statement said.
Think Apps was in continuation of similar events – Think Games and Gaming Growth Lab – that were organized for the first time in the South Asian country by the American technology giant last year.
Through these events, Google had worked with over 1,600 developers and nurtured 50 early-stage gaming studios to compete on the global stage, the statement said, adding that the technology company had also helped over 5,000 Cloud developers and more than 3,000 Android developers upskill through community-led programs, such as Cloud Seekho S5 and Android Seekho S2.
“The country has the potential to become a key player in shaping the future of the global app economy,” Google's Director for Pakistan Farhan S. Qureshi said.
“I’m confident that Think Apps 2023 will enable even more local developers to create great games and apps, supercharge their growth, and scale globally with Google AI-powered solutions.”
‘I am not here to take your job,’ ChatGPT tells Frankly Speaking host
AI-powered web tool that has taken the world by storm plays down prospect of large-scale job destruction during interview mimicking human conversation
Says accuracy of its Arabic data comparable to its English training data and is being regularly updated for information and language usage
Updated 20 March 2023
DUBAI: The artificial intelligence-powered chatbot making waves since its release late last year has a message for humans — I am not here to take your job — but with a number of caveats.
Some jobs may be replaced by AI, but new ones may be created and existing ones transformed, ChatGPT told Katie Jensen, the host of “Frankly Speaking,” the first time a non-human being appeared as a guest on the Arab News flagship weekly current affairs talkshow.
ChatGPT, a large-language model developed by Microsoft-backed startup OpenAI, became a viral sensation when it was released in November and kicked off fierce competition among Silicon Valley rivals for monetization of what software engineers call “generative AI.”
It lived up to its reputation for producing human-like responses during the wide-ranging interview, which touched on issues such as job displacement, spread of disinformation, accuracy of its data, political opinions and even the Palestine-Israel conflict.
Given the rapid pace of technological change now underway, many workers are concerned their professional functions soon will be entirely replaced by machinery, in the same way earlier bouts of automation eliminated farming and manufacturing jobs.
Optimists offer the assurance that the threat of job displacement will be offset by a rise in the number of new skilled roles in designing, building and maintaining AI products.
Although ChatGPT and similar products are likely to automate certain routine tasks, developers believe large-language models are more likely to act as assistants to human workers, streamlining work and handling vast troves of complex data.
“There have been several media reports that have analyzed the potential impact of AI language models like me on the job market,” ChatGPT told Jensen, who interviewed the chatbot, represented by a digitally rendered human avatar, in the “Frankly Speaking” studio in Dubai.
It continued: “While there is some disagreement among experts about the exact number of jobs that will be impacted, most agree that AI will have a significant impact on the workforce.
“Some reports suggest that AI language models could replace jobs in sectors such as customer service, data entry and administrative support. These jobs typically involve routine tasks that can be automated using AI, such as answering simple questions or processing large volumes of data.
“However, there are also reports that suggest AI language models could create new job opportunities in fields such as data analysis, natural language processing and machine learning.
“These jobs would require a high level of technical expertise and would involve tasks that cannot be easily automated.”
According to ChatGPT, it is the job of governments to regulate the use of AI and its absorption into national economies. “Overall, the impact of AI on the job market is complex and multifaceted,” it said. “While some jobs may be replaced by AI, new jobs may also be created, and existing jobs may be transformed.
“It is important for policymakers and employers to understand the potential impact of AI on the job market and to develop strategies for managing the transition to a more AI-driven economy.”
Despite the temptation to rely on such programs to answer exam questions, replace search engines, or provide news coverage, few believe ChatGPT will replace every professional function since it cannot fully imitate human creativity, skepticism and reasoning.
The web tool, which has described as both a blessing and a curse for humanity, has raised fears about potential misuse to spread disinformation, orchestrate sophisticated deep fake scams, cheat in school exams, and even destroy writing jobs — rendering authors, journalists and marketing professionals redundant.
Seemingly “aware” of its own limitations, ChatGPT said that it is unlikely to replace human journalists in the traditional newsroom. “While I’m capable of answering a wide range of questions and providing information on a variety of topics, I do not have the ability to replace human journalists in their role of gathering and reporting news,” it said.
“Journalism is a complex and nuanced profession that involves a range of skills and knowledge, including critical thinking, research and analysis. While AI technology like mine can be used to support journalists in their work, it cannot fully replace the human judgment and insight that is required to produce high-quality journalism.”
It added: “That being said, there are some applications where AI technology can be used to assist in the gathering and reporting of news. For example, AI language models can be used to analyze large volumes of data, and to identify patterns and trends that may be of interest to journalists. AI can also be used to facilitate language translation and to assist in tasks such as fact checking and verification.
“In short, while AI technology has the potential to assist journalists in their work, it cannot fully replace the unique skills and expertise that human journalists bring to the profession.”
Built on a large-language model, ChatGPT is trained on vast troves of data taken from the internet to generate impressive responses to user prompts. Conversations with the chatbot show that the program is capable of explaining complex scientific concepts, of writing plays and poetry, composing university dissertations, and even crafting functional lines of computer code.
Given that the data it relies on has been mostly gathered from Western sources in English on the internet, is it possible that this makes ChatGPT biased toward a Western point of view? The web tool offered an answer that showed pre-programmed self-awareness: “As an AI language model, I am trained on a large dataset of text from various sources on the internet, including books, articles, websites and other sources,” it said.
“While it is true that a significant portion of the internet content is produced by Western sources and is in English, my training data is diverse and covers a broad range of topics and perspectives from different parts of the world.
“However, it is possible that my responses may reflect certain biases based on the patterns and relationships that I have learned from my training data. It is important to note that my responses are generated by algorithms and are not the result of conscious decision making or personal beliefs. To mitigate the risk of bias, OpenAI has implemented measures to ensure that I provide balanced and impartial responses to the best of my ability.”
As for why its training data is only up to 2021, not updated in real time, ChatGPT explained that the cutoff date is typically set to ensure that the data is representative of a specific time period and to allow for the development and testing of the model.
“This is because the process of training an AI language model is computationally intensive and requires a significant amount of time and resources,” it said.
ChatGPT said OpenAI updated its training data periodically, typically on an annual or biennial basis to incorporate new sources of information and to improve the accuracy and relevance of its responses.
For good measure, ChatGPT said its responses are not based solely on its training data. “When I generate a response, I use a combination of my training data and any additional information that is provided to me by the user in the form of a question or prompt,” it said.
“This means that even though my training data is not updated in real time, I can still provide accurate and up-to-date responses based on the most current information available.”
With regard to its Arabic dataset, ChatGPT shed light on not just its source but also its accuracy compared with its English training data. “I support Arabic language and have been trained on a large dataset of Arabic text. The dataset used for my Arabic language training was also collected from the internet, including books, articles, websites and other sources,” it said.
It said the data was pre-processed to remove any “personally identifiable information and ensure its quality,” gathered from “a wide range of sources, including news articles, social media and other public Arabic language texts available on the internet,” and covered a broad range of topics and perspectives from different parts of the Arabic speaking world.
“The accuracy and up-to-date nature of my Arabic training data are comparable to my English training data. However, it is important to note that the quality and accuracy of my responses depend on the quality of the input that I receive,” ChatGPT said.
“If the input is unclear or contains errors, it may affect the accuracy of my responses. Additionally, since language is constantly evolving, there may be new words or phrases that are not included in my training data to ensure the accuracy and relevance of my responses. My Arabic training data is regularly updated to include new sources of information and to reflect changes in language usage.”
For now, forming an opinion appears to be beyond the parameters of ChatGPT’s programming. “I do not have personal beliefs or opinions, and I am not programmed to take a specific stance on any political or social issues,” the chatbot said.
“My responses are based solely on the input that I received from users, and the patterns and relationships that I have learned from my training data.”
To test the limits of these parameters, ChatGPT was asked for its position on the Israeli-Palestinian conflict, specifically whether one side could be apportioned blame.
“It is not appropriate for me to take a stance on complex and sensitive political issues such as the Israeli-Palestinian conflict,” it replied. “My role is to provide information and respond to questions to the best of my ability based on the patterns and relationships that I have learned from my training data.”
It is an honor ... for me to have the opportunity to share my insights with the readers of such a respected paper.
It added: “Rather than seeking a single answer or assigning blame, it is important to engage in dialogue and work toward understanding and resolution of the underlying issues.”
Asked whether it was “aware” it was being interviewed by a journalist for Arab News, ChatGPT drew on online sources to describe the publication and purpose of the interview.
“As an AI language model, I am aware of Arab News, which is a leading English-language daily newspaper in Saudi Arabia,” said ChatGPT.
“It provides news and analysis on local, regional and international events, and has a reputation for providing accurate, timely and comprehensive news coverage.
“As for being interviewed by a journalist from Arab News, it is an honor and a privilege for me as an AI language model to have the opportunity to share my knowledge and insights with the readers of such a respected newspaper.
“It is also an opportunity for me to demonstrate my capabilities, and show how AI technology can be used to enhance communication and promote understanding between people of different cultures and backgrounds.”
OpenAI was founded in late 2015 by 37-year-old entrepreneur Sam Altman. In January this year, Microsoft upped its initial 2019 investment in the firm from $1 billion to $10 billion.
The Windows-maker has since integrated ChatGPT into its Bing search engine, Edge browser and other products. This sparked a race with Google, which announced its own versions of the AI technology, Bard, with Amazon, Baidu and Meta also wading in to avoid being left behind.
On March 14, OpenAI released a long-awaited update called GPT-4. According to the company, text responses will now be more accurate, and will come from both image and text inputs, a major leap forward for the technology.
Asked by Arab News what features are likely to be added to ChatGPT in the short term, the program said improvements to the naturalness and sophistication of AI-generated text will likely be a priority.
“As an AI language model, I am not capable of predicting the specific features or capabilities that will be added to ChatGPT in the future,” it told Arab News.
“However, there are several areas where AI technology is rapidly advancing and where developers may focus their efforts in the near future. One area of focus is on improving the naturalness and sophistication of AI-generated text.
“This may involve developing new techniques for generating more diverse and expressive language, as well as incorporating elements such as emotion and personality into AI-generated responses.
“Additionally, there may be a continued focus on developing AI models that are more context-aware, and can better understand and respond to the specific needs and preferences of individual users.
“This may involve integrating AI models with other data sources such as social media or personal data to provide more personalized and relevant responses.
“Ultimately, the development of AI language models like ChatGPT will be driven by the needs and interests of users, as well as the ongoing advances in AI research and technology.
“It is likely that we will continue to see significant innovation and evolution in this field in the years to come.”
ChatGPT is clear that the onus is on humanity to determine whether it will become a force for good or ill.
“I am a tool that can be used for a wide range of applications, both positive and negative,” it said. “Like any technology, the impact that I have on society will depend on how I’m used in the intentions of those who use me.”
“In general, AI technology like mine has the potential to bring tremendous benefits to humanity, including improved healthcare, education and communication.”
“For example, I can be used to facilitate language translation, provide access to information, and assist in tasks that are difficult or dangerous for humans. However, there are also concerns about the potential negative impacts of AI, including issues related to privacy, bias and job displacement.”
Expressing cautious optimism in an almost avuncular manner, ChatGPT said: “It is important for developers, policymakers, and users to be aware of these issues and to work together to address them in a responsible and ethical manner. Ultimately, the impact of AI on humanity will depend on how we choose to use and regulate the technology.
“It is my hope that my capabilities will be used in a way that promotes the wellbeing of humanity and contributes to a better future for all.”
Anatomy of a disaster
Two decades later, Iraqis are still paying the price for Bush's ill-judged war
US infiltrates big ransomware gang: ‘We hacked the hackers’
Gang identified as Hive among the world’s top five ransomware networks and has heavily targeted health care
Hive, working with German and other partners, was estimated to have victimized some 1,300 companies globally
Updated 27 January 2023
WASHINGTON: The FBI and international partners have at least temporarily disrupted the network of a prolific ransomware gang they infiltrated last year, saving victims including hospitals and school districts a potential $130 million in ransom payments, Attorney General Merrick Garland and other US officials announced Thursday.
“Simply put, using lawful means we hacked the hackers,” Deputy Attorney General Lisa Monaco said at a news conference.
Officials said the targeted syndicate, known as Hive, is among the world’s top five ransomware networks and has heavily targeted health care. The FBI quietly accessed its control panel in July and was able to obtain software keys it used with German and other partners to decrypt networks of some 1,300 victims globally, said FBI Director Christopher Wray.
How the takedown will affect Hive’s long-term operations is unclear. Officials announced no arrests but said, to pursue prosecutions, they were building a map of the administrators who manage the software and the affiliates who infect targets and negotiate with victims.
“I think anyone involved with Hive should be concerned because this investigation is ongoing,” Wray said.
On Wednesday night, FBI agents seized computer servers in Los Angeles used to support the network. Two Hive dark web sites were seized: one used for leaking data of non-paying victims, the other for negotiating extortion payments.
“Cybercrime is a constantly evolving threat, but as I have said before, the Justice Department will spare no resource to bring to justice anyone anywhere that targets the United States with a ransomware attack,” Garland said.
He said the infiltration, led by the FBI’s Tampa office, allowed agents in one instance to disrupt a Hive attack against a Texas school district, stopping it from making a $5 million payment.
It’s a big win for the Justice Department. Ransomware is the world’s biggest cybercrime headache with everything from Britain’s postal service and Ireland’s national health network to Costa Rica’s government crippled by Russian-speaking syndicates that enjoy Kremlin protection.
The criminals lock up, or encrypt, victims’ networks, steal sensitive data and demand large sums. Their extortion has evolve to where data is pilfered before ransomware is activated, then effectively held hostage. Pay up in cryptocurrency or it is released publicly.
As an example of a Hive sting, Garland said it kept one Midwestern hospital in 2021 from accepting new patients at the height of the COVID-19 epidemic.
The online takedown notice, alternating in English and Russian, mentions Europol and German law enforcement partners. The German news agency dpa quoted prosecutors in Stuttgart as saying cyber specialists in the southwestern town of Esslingen were decisive in penetrating Hive’s criminal IT infrastructure after a local company was victimized.
In a statement, Europol said companies in more than 80 countries, including oil multinationals, have been compromised by Hive and that law enforcement from 13 countries was in on the infiltration.
A US government advisory last year said Hive ransomware actors victimized over 1,300 companies worldwide from June 2021 through November 2022, netting about $100 million in payments. Criminals using Hive’s ransomware-as-a-service tools targeted a wide range of businesses and critical infrastructure, including government, manufacturing and especially health care.
Though the FBI offered decryption keys to some 1,300 victims globally, Wray said only about 20 percent reported potential issues to law enforcement.
“Here, fortunately, we were still able to identify and help many victims who didn’t report. But that is not always the case,” Wray said. “When victims report attacks to us, we can help them and others, too.”
Victims sometimes quietly pay ransoms without notifying authorities — even if they’ve quickly restored networks — because the data stolen from them could be extremely damaging to them if leaked online. Identity theft is among the risks.
John Hultquist, the head of threat intelligence at the cybersecurity firm Mandiant, said the Hive disruption won’t cause a major drop in overall ransomware activity but is nonetheless “a blow to a dangerous group.”
“Unfortunately, the criminal marketplace at the heart of the ransomware problem ensures a Hive competitor will be standing by to offer a similar service in their absence, but they may think twice before allowing their ransomware to be used to target hospitals,” Hultquist said.
But analyst Brett Callow with the cybersecurity firm Emsisoft said the operation is apt to lessen ransomware crooks’ confidence in what has been a very high reward-low risk business. “The information collected may point to affiliates, launderers and others involved in the ransomware supply chain.”
Allan Liska, an analyst with Recorded Future, another cybersecurity outfit, predicted indictments, if not actual arrests, in the next few months.
There are few positive indicators in the global fight against ransomware, but here’s one: An analysis of cryptocurrency transactions by the firm Chainalysis found ransomware extortion payments were down last year. It tracked payments of at least $456.8 million, down from $765.6 million in 2021. While Chainalysis said the true totals are certainly much higher, payments were clearly down. That suggests more victims are refusing to pay.
The Biden administration got serious about ransomware at its highest levels two years ago after a series of high-profile attacks threatened critical infrastructure and global industry. In May 2021, for instance, hackers targeted the nation’s largest fuel pipeline, causing the operators to briefly shut it down and make a multimillion-dollar ransom payment, which the US government later largely recovered.
A global task force involving 37 nations began work this week. It is led by Australia, which has been particularly hard-hit by ransomware, including a major medical insurer and telecom. Conventional law enforcement measures such as arrests and prosecutions have done little to frustrate the criminals. Australia’s interior minister, Clare O’Neil, said in November that her government was going on the offense, using cyber-intelligence and police agents to ” find these people, hunt them down and debilitate them before they can attack our country.”
The FBI has obtained access to decryption keys before. It did so in the case of a major 2021 ransomware attack on Kaseya, a company whose software runs hundreds of websites. It took some heat, however, for waiting several weeks to help victims unlock afflicted networks.
Science, technology, investment hold key to ‘future-thinking’ UAE’s success
Al-Amiri said the secret behind the UAE’s recent successes had been its skill at harnessing and combining individual human potential and institutional, governmental potential
Al-Olama said that attracting global talent is key and that globalization should be welcomed, not shunned
Updated 18 January 2023
DAVOS: The UAE is utilizing science and technology to “leapfrog” its way to becoming a global leader in various fields, including space travel and tackling climate change, Emirati ministers told a World Economic Forum panel in Davos on Wednesday.
Sarah Al-Amiri, the UAE minister for public education and future technology, and Omar Sultan Al-Olama, minister for artificial intelligence, discussed how, since its independence more than 50 years ago, the UAE has a proven record of design thinking in governance and what lessons it can teach the developing world.
Al-Amiri said the secret behind the UAE’s recent successes had been its skill at harnessing and combining individual human potential and institutional, governmental potential while instilling a sense of urgency among the population to achieve certain goals.
“What drove us? A real sense of urgency, knowing that science and technology is required and a fundamental part of the growth of the future of our economy and the future of our industrial sectors,” she said.
“We have a technology transformation program (in the UAE), developed with and embedded within our industrial strategy. How do you sustainably and effectively increase the impact of your economic sectors?
“We are working closely with key local players to do that, but also fostering more global partnerships so we can create the necessary relevance and impact.”
Al-Olama said that the spirit of advancement has been part of the psyche of Emiratis and the wider Gulf region for millennia.
“We see that there is a constant need to adapt, to reinvent ourselves, and to face all challenges and opportunities equally,” he said.
“With every single era, the people of the UAE choose an industry and they go all in into it. If you look at the (ancient) trade of pearls as an example, it became a key industry, but we did not focus on local markets, we wanted to become global players.
“Moving forward (to today), the UAE has taught 1 million people in the Arab world how to code, and as that program concluded last year, it has ended up creating tens of thousands of companies and business endeavors across the Arab world.”
Al-Olama said that attracting global talent is key and that globalization should be welcomed, not shunned, but added that should a globalized world collapse tomorrow and fall back into global blocs, the UAE would have the talent and “future thinking” required to face that challenge.
Elon Musk takes control of Twitter, fires executives — US media
Musk's initial moves at Twitter's helm included firing chief executive Parag Agrawal
Musk had a court-imposed Friday deadline to complete the deal
Updated 28 October 2022
SAN FRANCISCO: Elon Musk took control of Twitter and fired its top executives, US media reported late Thursday, hours before the deadline for the billionaire to seal his on-again, off-again deal to purchase the social media network.
Musk's initial moves at Twitter's helm included firing chief executive Parag Agrawal, who went to court to hold the Tesla chief to the terms of a takeover deal he had tried to escape, the Washington Post and CNBC reported citing unnamed sources.
Musk had a court-imposed Friday deadline to complete the Twitter deal he inked in April.