Agregátor RSS
Vivo X300 Ultra je v prodeji. Při nákupu profesionální fotosady můžete získat bonusy za více než 13 tisíc
Reporter’s notebook: In Nepal and Sri Lanka, AI boom brings hope
The soap refill dispenser at a cafe in TRACE Expert City — a technology hub in Colombo, Sri Lanka — boldly declares that it was delivered by ”USAID, from the American people.”
The device is a relic from the past, reflecting goodwill that once existed between the US and Sri Lanka. Now, as external aid through USAID and other entities dries up, the island nation is looking inward for its future.
You’ll find some of that focus at TRACE’s campus, where energetic AI entrepreneurs with international ambitions brim with fresh ideas. Startup Jendo Innovations is already deploying AI-driven medical technology internationally. “We have the best affordable talent in the region to develop applications on AI,” said Heminda Jayaweera, executive director of TRACE.
Meanwhile, in Kathmandu, Nepal, venture capitalist Preeti Adhikary and medical doctor Kisu Rawal beam with joy as women entrepreneurs wrap up a series of AI training sessions. Adhikary sees in the women newfound confidence and fresh ideas about how to expand and streamline their businesses.
While much of the chatter about AI in the US carries with it an element of danger — hallucinations, rogue AI agents, leaked or stolen data — in Nepal the conversation centers around empowerment and upward mobility.
“While AI has been an equalizer…, it’s on us to find underserved communities and demographics to educate and enable,” said Adhikary, who is a general partner at Momo VC. She works to connect the Nepal tech ecosystem through initiatives like The Great Nepali Diaspora, a global network of Nepali talent, and The Empowered Women Network, which organized the recent AI training session.
Preeti Adhikary, venture capitalist and a general partner at Momo VC.
The Empowered Women Network
To be sure, AI is still in the Wild West phase in Nepal, Sri Lanka and many developing economies — just as it is elsewhere in the world. Locally, some startups have focused AI tools on their unique cultures, demographics and business environments.
Some are focused on AI for good, especially in rural areas, where 33% of the population resides, according to the World Bank.
Mobile devices are ubiquitous in those areas, which opens the door for AI-driven healthcare that can quickly diagnose diseases and reduce long lines in remote hospitals.
But culture also presents unique data challenges. Literacy remains an issue, and locals prefer interacting with AI using voice-driven methods rather than text, which changes customer feedback and service flows.
Product recommendations are also made by word-of-mouth, cutting off one layer of product discovery and reshaping how AI models are built.
Because workforces in Nepal and Sri Lanka are still emerging, the concern over AI-driven job losses is not as dire as in the US, where job cuts have been attributed to automation via AI. And unlike in the US, legacy systems and technical debt are not a burden here.
Enterprises with lean IT budgets have little appetite for digitizing legacy data, which could slow AI adoption. So, bootstrapped startups, facing limited hardware, capital and talent, hope to make their workflows and processes AI-native from the start.
Even so, the lack of resources makes innovation and skills development difficult. Nvidia GPUs, for instance, are particularly scarce in Nepal and Sri Lanka and prohibitively expensive.
To work around the challenges, startups have turned to small language models (SLMs) and popular AI tools. “AI tools — which are powerful even at the free version — and open-source models have meant AI is truly democratized now,” Adhikary said.
AI adoption is poised to grow in both Nepal and Sri Lanka, both of which have Gen-Z-inspired governments that grew up on social media and are AI-ready. Nepal last year published a national AI strategy outlining plans to boost infrastructure, skills readiness and countrywide awareness.
It has room to grow. Nepal ranks 106 out of 190 countries in Oxford Insights’ 2025 Government AI Readiness Index. The nation’s goal is to reach the top 50.
Meanwhile, Sri Lanka is ranked 80th and is pushing AI adoption. The country in 2024 established the Ministry of Digital Economy and has an AI strategy to boost the local ecosystem.
And Sri Lanka’s new digital nomad program could bring diaspora and foreign tech workers back onshore to help bolster local AI development and adoption.
As in other parts of the world, there is an ongoing need for mentorship, domain expertise and access to networks to grow and scale, according to Adhikary. But AI entrepreneurs in Nepal recognize the benefits of collaboration.
“I’m pleasantly surprised … that a nation with a newish tech ecosystem can compete with New York and Silicon Valley in terms of ideas and technical acumen,” Adhikary said.
Industries Most Exposed to AI Are Not Only Seeing Productivity Gains but Jobs and Wage Growth Too
New technologies rarely leave work untouched. They also rarely eliminate the need for human contribution altogether.
Forecasts of the impact of artificial intelligence range from the apocalyptic to the utopian. An October 2025 report from Senate Democrats, for example, predicted AI will destroy millions of US jobs. A couple of years earlier, consultant company McKinsey forecast AI will add trillions to the global economy, while emphasizing job losses can be mitigated by training workers to do new things.
The problem is that many of these claims are based on projections, overly simplified surveys, or thought experiments rather than observed changes in the economy. That makes it hard for the public, and often policymakers, to know what to trust.
As a labor economist who studies how technology and organizational change affect productivity and well-being, I believe a better place to start is with actual data on output, employment, and wages—which are all looking relatively more hopeful.
AI and JobsIn one of my new research papers with economist Andrew Johnston, we studied how exposure to generative AI affected industries across America between 2017 and 2024, using administrative data that covers nearly all employers. Our analysis covered a crucial period when generative AI use exploded, allowing us to analyze the effect within businesses and industries.
We measured AI exposure using occupation-level task data matched to each industry and state’s occupational workforce mix prior to the pandemic. A state and industry with more workers in roles requiring language processing, coding, or data tasks scored higher on exposure, for example, compared with one with more plumbers and electricians.
We then took that exposure ranking by occupation and looked at changes in the standard deviation in occupational exposure, comparing that with labor market and GDP across states and industries from 2017 to 2024.
Think of a standard deviation as roughly the gap between a paramedic—whose work centers on physical assessment, emergency response, and hands-on care that AI cannot easily replicate—and a public relations manager, whose work involves drafting communications, analyzing sentiment, and synthesizing information that AI tools handle well. That gap in AI exposure is roughly what we’re measuring when we ask: Does being on the higher-exposure side of that divide change your industry’s trajectory?
This data allowed us to answer two questions: When AI tools became widely available following the public release of ChatGPT in late 2022, did states and industries that were more exposed to generative AI become more productive, and what happened to workers?
Our answers are more encouraging, and more nuanced, than much of the public debate suggests.
We found that industries in states that were more exposed to AI experienced faster productivity growth beginning in 2021—before ChatGPT reached the public—driven by enterprise tools already embedded in professional workflows, including GitHub Copilot for software development, Jasper for marketing and content writing, and Microsoft’s GPT-3-powered business applications. In 2024, for example, industries whose AI exposure was one standard deviation higher saw gains of 10% in productivity, 3.9% in jobs, and 4.8% in wages than comparable industries in the same state.
Those patterns suggest that, at least so far, AI has acted as a productivity-enhancing tool that boosts employment and wages rather than a simple substitute for labor.
Augmentation Versus DisplacementA crucial distinction in the data is between tasks where AI works with people and tasks where AI can act more independently. In sectors where AI mainly complements workers—think marketing, writing, or financial analysis—our data show that employment rose by about 3.6% per standard deviation increase in exposure.
In sectors where AI can execute tasks more autonomously—including basic data processing, generating boilerplate code, or handling standardized customer interactions—we found no significant employment change, though workers in those roles saw slower wage growth.
What these findings suggest is that when AI lowers the cost of completing a task and raises worker productivity, companies expand output enough to increase their demand for labor overall—the same logic that explains why power tools didn’t eliminate construction workers.
The economic question is not whether any given task disappears. It is whether businesses and workers can reorganize fast enough to create new productive combinations. And so far, in most sectors, our evidence suggests they can.
But state policies also matter: These benefits were concentrated in the states with more efficient labor markets, meaning that the impact of generative AI on workers and the economy also depends on the types of policies and institutions of the local economy.
Importantly, these findings hold beyond occupational exposure. In additional work with co-authors at the Bureau of Economic Analysis, we found a similar effect on GDP and employment when looking at actual AI utilization—that is, how often workers use AI. Drawing on the Gallup Workforce Panel, we measured workers actively using AI daily or multiple times a week. We found that each percentage-point increase in the share of frequent AI users in a state and industry is associated with roughly 0.1% to 0.2% higher real output and 0.2% to 0.4% higher employment.
To put that in context: The share of frequent AI users across all occupations rose from about 12% in mid-2024 to 26% by late 2025, a shift our estimates suggest corresponds to roughly 1.4% to 2.8% higher real output—or about 1 to 2 percentage points of annualized growth over that period.
New technologies rarely leave work untouched. But they also rarely eliminate the need for human contribution altogether. Instead, they change the composition of work, as our research shows. Some tasks shrink. Others expand. New ones emerge that were previously too costly or too hard to perform at scale. Put simply, some occupations might go away, but most of them just change.
If anything, the trends documented here are likely to strengthen rather than fade. Not only are generative AI tools rapidly improving, but also the experimentation and research and development that many workers and companies are engaging in are likely to pay large dividends. These investments—often referred to as intangible capital—tend to get unlocked a few years after a technology comes onto the scene, once complementary investments have been made.
The Role of Companies and ManagersWhether AI leads to anxiety or adaptation for workers depends in part on what happens inside organizations. Using additional data collected over many years in the Gallup Workforce Panel covering more than 30,000 US employees from 2023 to 2026, I found in a 2026 paper that workplace adoption of generative AI rose quickly over the period, with the share of workers using AI often increasing from 9% to 26%.
But the more important finding is that adoption was far more common where workers believed their organization had communicated a clear AI strategy and where employees said they trust leadership. This suggests that growing adoption and effective use of AI depends not only on the availability of the technology but on whether managers make its use clear, credible, and safe.
Where that clarity exists, frequent AI use is associated with higher engagement and job satisfaction, and it even reverses the burnout penalties that appear elsewhere.
In other words, the broader economic effects of AI depend not only on how sophisticated the tools are but on whether companies and managers create environments where workers can experiment, reorganize tasks, and integrate new tools into productive routines. That is, if employees do not feel the psychological safety to experiment, they are less likely to use AI, and they are especially less likely to use it for higher-value work.
That is precisely the kind of adaptation that I believe makes labor markets more resilient than the most alarmist forecasts suggest.
This article is republished from The Conversation under a Creative Commons license. Read the original article.
The post Industries Most Exposed to AI Are Not Only Seeing Productivity Gains but Jobs and Wage Growth Too appeared first on SingularityHub.
Hackers exploit Marimo flaw to deploy NKAbuse malware from Hugging Face
When LKML Patches Signal Exploitation Risk Before CVE Assignment
Vybíráme nejlepší fotomobily současnosti. Tohle si kupte místo foťáku
Příběhy Kingdom Come rozšíří povídková kniha. Vrátí se Michal a David, dojde i na nové postavy
Google expands Gemini AI use to fight malicious ads on its platform
Americans who masterminded Nork IT worker fraud sentenced to 200 months behind bars
Two Americans have been jailed for a combined 200 months for helping North Korea generate $5 million through fraudulent IT worker schemes.…
Gemini opustilo prohlížeč a je rovnou na ploše. Google na Macy vpustil AI
New ATHR vishing platform uses AI voice agents for automated attacks
Most "AI SOCs" Are Just Faster Triage. That's Not Enough.
Nová CAPTCHA na katastru nemovitostí
Lego leze do peněz. Tady je 15 nejdražších setů, které jsou aktuálně v prodeji. Začínáme nad 11 tisíci
Rust 1.95.0
Thunderbolt, open-source AI klient od Mozilly
ThreatsDay Bulletin: Defender 0-Day, SonicWall Brute-Force, 17-Year-Old Excel RCE and 15 More Stories
Git identity spoof fools Claude into giving bad code the nod
Security boffins say Anthropic's Claude can be tricked into approving malicious code with just two Git commands by spoofing a trusted developer's identity.…
4K projektor za hubičku. Kvalitní Optoma Photo Life PK31 zlevnil na 11 tisíc
Microsoft’s Windows Recall still allows silent data extraction
Microsoft’s Windows Recall feature remains vulnerable to complete data extraction despite a major security overhaul, according to a cybersecurity researcher who says malware running in a user’s context can quietly siphon off everything Recall has captured, without administrator privileges, kernel exploits, or breaking encryption.
Alexander Hagenah, executive director at Zürich-based financial infrastructure operator SIX Group, made the claim in a LinkedIn post, where he also published a proof-of-concept tool called TotalRecall Reloaded to demonstrate the issue.
Hagenah first exposed Recall’s security flaws in 2024, forcing Microsoft to pull the feature from preview and rebuild it. Microsoft relaunched Recall in April 2025, saying the new architecture would restrict “attempts by latent malware trying to ‘ride along’ with a user authentication to steal data.” Hagenah said it does not.
“When you use Recall normally, TotalRecall Reloaded silently holds the door open behind you and then extracts what Recall has ever captured. That is precisely the scenario Microsoft’s architecture is supposed to restrict,” he wrote in the post.
Hagenah wrote in the post that he disclosed the research to Microsoft’s Security Response Center on March 6, submitting full source code and reproduction steps. Microsoft reviewed the case for a month and closed it on April 3, telling him the behavior “does not represent a bypass of a security boundary or unauthorized access to data.”
“Microsoft says this is by design,” Hagenah wrote. “That worries me.”
In an email response to CSO, a Microsoft spokesperson said, “After careful investigation, we determined that the access patterns demonstrated are consistent with intended protections and existing controls, and do not represent a bypass of a security boundary or unauthorized access to data. The authorization period has a timeout and anti-hammering protection that limit the impact of malicious queries.”
Hagenah’s research does not challenge Microsoft’s encryption, which he said is sound. The gap, he told CSO, is in how decrypted data is handled once it leaves the enclave.
“Plaintext screenshots and extracted text end up in an unprotected process for display,” he told CSO. “As long as decrypted content crosses into a process that same-user code can access, someone will find a way in.”
What a fix would requireA fix is technically feasible, Hagenah said.
“The short-term fix is fairly straightforward. Microsoft could add stronger code integrity and process protections to AIXHost.exe, the process that renders the Recall timeline. Right now, it has none, which makes the injection path possible. That would block the specific technique I demonstrated and materially raise the bar,” he said.
The longer-term problem runs deeper, he said. “Microsoft should rethink how decrypted data is handled after it leaves the enclave. The cryptography and enclave design are genuinely well done, and I want to be clear about that. The problem is that plaintext screenshots and extracted text end up in an unprotected process for display. As long as decrypted content crosses into a process that same-user code can access, someone will find a way in,” he said.
“A durable fix would mean either rendering inside a protected process or adopting a compositing model where raw data never leaves the trust boundary. That is a bigger effort, but it is the only way to close this class of issue properly,” he said.
Exploitation riskThe barrier to weaponizing this technique is lower than Microsoft’s security messaging would suggest, Hagenah said.
“They only need code running in the user’s context and a way to reuse the authorized Recall session,” he said. “That is a much lower bar than many people would assume from Microsoft’s security messaging.”
While Recall’s limitation to Copilot+ PCs and its opt-in status reduce the scale of exposure, targeted abuse is a realistic near-term risk, he said. “For targeted abuse, surveillance, or high-value user collection, this is absolutely realistic,” he said.
Hagenah said he published the source code deliberately so defenders, EDR vendors, and security teams could build detections before threat actors operationalize the technique independently. “In my view, that gives the defensive side a valuable head start,” he said.
Independent security researcher Kevin Beaumont reached a similar conclusion after separately testing the current Recall implementation. “Yep, you can just read the database as a user process,” Beaumont wrote on Mastodon on March 11. “The database also contains all manner of fields that aren’t publicly disclosed for tracking the user’s activity. No AV or EDR alerts triggered,” he wrote.
The article originally appeared in CSO.
- « první
- ‹ předchozí
- …
- 35
- 36
- 37
- 38
- 39
- 40
- 41
- 42
- 43
- …
- následující ›
- poslední »



