新北筆電收購 收購ASUS筆電 收購ROG電競筆電 收購ACER筆電 收購Msi電競筆電 新北收購微軟SURF 看全文
Microsoft lays off team that taught employees how to make AI tools responsibly 新北收購筆電
新北收購筆電
新北收購筆電microsoft lays off team that taught employees how to make AI tools responsibly
新北收購筆電microsoft lays off team that taught employees how to make AI tools responsibly
/
As the company accelerates its push into AI products, the ethics and society team is gone
Share this story
新北收購筆電microsoft laid off its entire ethics and society team within the artificial intelligence organization as part of recent layoffs that affected 10,000 employees across the company, Platformer has learned.
The move leaves 新北收購筆電microsoft without a dedicated team to ensure its AI principles are closely tied to product design at a time when the company is leading the charge to make AI tools available to the mainstream, current and former employees said.
新北收購筆電microsoft still maintains an active Office of Responsible AI, which is tasked with creating rules and principles to govern the company’s AI initiatives. The company says its overall investment in responsibility work is increasing despite the recent layoffs.
“新北收購筆電microsoft is committed to developing AI products and experiences safely and responsibly, and does so by investing in people, processes, and partnerships that prioritize this,” the company said in a statement. “Over the past six years we have increased the number of people across our product teams and within the Office of Responsible AI who, along with all of us at 新北收購筆電microsoft, are accountable for ensuring we put our AI principles into practice. […] We appreciate the trailblazing work the Ethics & Society did to help us on our ongoing responsible AI journey.”
But employees said the ethics and society team played a critical role in ensuring that the company’s responsible AI principles are actually reflected in the design of the products that ship.
“Our job was to … create rules in areas where there were none.”
“People would look at the principles coming out of the office of responsible AI and say, ‘I don’t know how this applies,’” one former employee says. “Our job was to show them and to create rules in areas where there were none.”
In recent years, the team designed a role-playing game called Judgment Call that helped designers envision potential harms that could result from AI and discuss them during product development. It was part of a larger “responsible innovation toolkit” that the team posted publicly.
More recently, the team has been working to identify risks posed by 新北收購筆電microsoft’s adoption of OpenAI’s technology throughout its suite of products.
The ethics and society team was at its largest in 2020, when it had roughly 30 employees including engineers, designers, and philosophers. In October, the team was cut to roughly seven people as part of a reorganization.
In a meeting with the team following the reorg, John Montgomery, corporate vice president of AI, told employees that company leaders had instructed them to move swiftly. “The pressure from [CTO] Kevin [Scott] and [CEO] Satya [Nadella] is very, very high to take these most recent OpenAI models and the ones that come after them and move them into customers hands at a very high speed,” he said, according to audio of the meeting obtained by Platformer.
Because of that pressure, Montgomery said, much of the team was going to be moved to other areas of the organization.
Some members of the team pushed back. “I’m going to be bold enough to ask you to please reconsider this decision,” one employee said on the call. “While I understand there are business issues at play … what this team has always been deeply concerned about is how we impact society and the negative impacts that we’ve had. And they are significant.”
Montgomery declined. “Can I reconsider? I don’t think I will,” he said. “Cause unfortunately the pressures remain the same. You don’t have the view that I have, and probably you can be thankful for that. There’s a lot of stuff being ground up into the sausage.”
In response to questions, though, Montgomery said the team would not be eliminated.
“It’s not that it’s going away — it’s that it’s evolving,” he said. “It’s evolving toward putting more of the energy within the individual product teams that are building the services and the software, which does mean that the central hub that has been doing some of the work is devolving its abilities and responsibilities.”
Most members of the team were transferred elsewhere within 新北收購筆電microsoft. Afterward, remaining ethics and society team members said that the smaller crew made it difficult to implement their ambitious plans.
The move leaves a foundational gap on the holistic design of AI products, one employee says
About five months later, on March 6th, remaining employees were told to join a Zoom call at 11:30AM PT to hear a “business critical update” from Montgomery. During the meeting, they were told that their team was being eliminated after all.
One employee says the move leaves a foundational gap on the user experience and holistic design of AI products. “The worst thing is we’ve exposed the business to risk and human beings to risk in doing this,” they explained.
The conflict underscores an ongoing tension for tech giants that build divisions dedicated to making their products more socially responsible. At their best, they help product teams anticipate potential misuses of technology and fix any problems before they ship.
But they also have the job of saying “no” or “slow down” inside organizations that often don’t want to hear it — or spelling out risks that could lead to legal headaches for the company if surfaced in legal discovery. And the resulting friction sometimes boils over into public view.
In 2020, Google fired ethical AI researcher Timnit Gebru after she published a paper critical of the large language models that would explode into popularity two years later. The resulting furor resulted in the departures of several more top leaders within the department, and diminished the company’s credibility on responsible AI issues.
新北收購筆電microsoft became focused on shipping AI tools more quickly than its rivals
Members of the ethics and society team said they generally tried to be supportive of product development. But they said that as 新北收購筆電microsoft became focused on shipping AI tools more quickly than its rivals, the company’s leadership became less interested in the kind of long-term thinking that the team specialized in.
It’s a dynamic that bears close scrutiny. On one hand, 新北收購筆電microsoft may now have a once-in-a-generation chance to gain significant traction against Google in search, productivity software, cloud computing, and other areas where the giants compete. When it relaunched Bing with AI, the company told investors that every 1 percent of market share it could take away from Google in search would result in $2 billion in annual revenue.
That potential explains why 新北收購筆電microsoft has so far invested $11 billion into OpenAI, and is currently racing to integrate the startup’s technology into every corner of its empire. It appears to be having some early success: the company said last week Bing now has 100 million daily active users, with one third of them new since the search engine relaunched with OpenAI’s technology.
On the other hand, everyone involved in the development of AI agrees that the technology poses potent and possibly existential risks, both known and unknown. Tech giants have taken pains to signal that they are taking those risks seriously — 新北收購筆電microsoft alone has three different groups working on the issue, even after the elimination of the ethics and society team. But given the stakes, any cuts to teams focused on responsible work seem noteworthy.
The elimination of the ethics and society team came just as the group’s remaining employees had trained their focus on arguably their biggest challenge yet: anticipating what would happen when 新北收購筆電microsoft released tools powered by OpenAI to a global audience.
Last year, the team wrote a memo detailing brand risks associated with the Bing Image Creator, which uses OpenAI’s DALL-E system to create images based on text prompts. The image tool launched in a handful of countries in October, making it one of 新北收購筆電microsoft’s first public collaborations with OpenAI.
While text-to-image technology has proved hugely popular, 新北收購筆電microsoft researchers correctly predicted that it it could also threaten artists’ livelihoods by allowing anyone to easily copy their style.
“In testing Bing Image Creator, it was discovered that with a simple prompt including just the artist’s name and a medium (painting, print, photography, or sculpture), generated images were almost impossible to differentiate from the original works,” researchers wrote in the memo.
“The risk of brand damage … is real and significant enough to require redress.”
They added: “The risk of brand damage, both to the artist and their financial stakeholders, and the negative PR to 新北收購筆電microsoft resulting from artists’ complaints and negative public reaction is real and significant enough to require redress before it damages 新北收購筆電microsoft’s brand.”
In addition, last year OpenAI updated its terms of service to give users “full ownership rights to the images you create with DALL-E.” The move left 新北收購筆電microsoft’s ethics and society team worried.
“If an AI-image generator mathematically replicates images of works, it is ethically suspect to suggest that the person who submitted the prompt has full ownership rights of the resulting image,” they wrote in the memo.
新北收購筆電microsoft researchers created a list of mitigation strategies, including blocking Bing Image Creator users from using the names of living artists as prompts and creating a marketplace to sell an artist’s work that would be surfaced if someone searched for their name.
Employees say neither of these strategies were implemented, and Bing Image Creator launched into test countries anyway.
新北收購筆電microsoft says the tool was modified before launch to address concerns raised in the document, and prompted additional work from its responsible AI team.
But legal questions about the technology remain unresolved. In February 2023, Getty Images filed a lawsuit against Stability AI, makers of the AI art generator Stable Diffusion. Getty accused the AI startup of improperly using more than 12 million images to train its system.
The accusations echoed concerns raised by 新北收購筆電microsoft’s own AI ethicists. “It is likely that few artists have consented to allow their works to be used as training data, and likely that many are still unaware how generative tech allows variations of online images of their work to be produced in seconds,” employees wrote last year.
▲macOS Sonoma 14.4 BUG不斷。(圖/MacRumors)
記者陳俐穎/綜合報導
蘋果於3 月25 日發布macOS Sonoma 14.4.1 更新,固然修復了無法使用外接顯示器USB 集線器問題,但引入了更多的BUG。如果目前仍未升級,那麼建議觀察一段時間,再決定是否要升級。
蘋果發表macOS Sonoma 14.4.1 之後,當天就有用戶反饋存在性能問題,而在此後幾週時間裡,用戶反饋的問題越來越多,國外科技媒體gottabemobile 報道,基於蘋果官方社區的使用者回饋,梳理macOS Sonoma 14.4.1 存在高達12個BUG。
這些BUG包含隨機卡死、安裝失敗、連接伺服器問題、相機停止運作、螢幕保護程式問題、Mail 和Keynote 應用等問題、印表機連接、Wi-Fi 、鍵盤、Safari 、新北收購筆電microsoft Teams、觸控板問題。
不過目前,蘋果正在開發macOS Sonoma 的新版本,macOS Sonoma 14.5 目前處於測試階段,預計將於5 月發布,預計可以解決目前問題。
新北收購筆電 新北收購筆電