fall25:  The AGENTIC OLYMPICS - Is Nvidia free to win this race or has wall street declared chiense walkover
MOTION AGAINST GETTING INTO DEBT CERTIFIED BY UNIVERSITIES 4 YEAR DEGREES
aka water water everywhere not drop to drink, intelligence everywhere not a trust map to link Most exciting time (July update) to be alive- supercomputer 1/7 thks Memphis! (more) ..why chat revolution of 2022 may have been by itself the least important of West Coast intel leaps every 3 years of 21stC
English Language Model- purpose to CODE trust and productive intelligences of millennials everywhere. 275 years of artificial cases from USA; 103 years from Konisberg Russia. Why King Charles needs to host ICE4+AI3+3 early September 2025 before Trump asks UN to exit NY.
Sub-ED: .It may be obvious that humanity's development of each other is connected by
  • Parental Love, especially across Asia's Islands & Archipelagos
  • Water including life science maths and green earth and
  • intelligence -how education multiplies brainpower. But intelligence @2025 is particularly curious driven by 10**18 more tech in last 60 years; since 2010 we've seen million fold more impact of satellites and computers :part 2 of what some call artificial intelligence); again from 1995 satellite acceleration of webs evolved borderless sharing of life critical knowhow through million fold human data-mapping devices including phone, text, camera uniting all human senses and steve jobs university in a phone; earlier Moores law's engineering of chips on both sides of Pacific delivered 1000 fold more tech 65-80 and another 1000 fold from 1980-95
    DO WE ALL LOVE TAIWAN as much as AI20s supercomputing & neural net wizards such as Jensen Huang, Demis Hassabis, Yann Lecun ? Perplexity explains why so few people linking to 20 million people leading every agency of AI that educational futures revolve round:No other small or island nation is currently aiming to train as many young AI professionals, relative to its population, as Taiwan—though Singapore, Hong Kong and Israel remain the benchmarks for workforce concentration123. In short: Taiwan’s AI talent drive is among the world’s most ambitious for its size, and it is on track to join or even surpass the global leaders in AI talent concentration in the coming years.Economic Impact: AI is projected to deliver over TWD 3.2 trillion (USD 101.3 billion) in economic benefits to Taiwan by 2030—more than 13% of current GDP. In 2023 alone, Google’s AI-related activities contributed TWD 682.2 billion and supported nearly 200,000 jobs in Taiwan3
  • HUMANITY & INTELLIGENCE's FUTURE
    Thanks to Jensen Huang the last decade has been most exciting of 75 years dad Norman Macrae 1:: 2 and then I have had privilege to question inteliligence's future. In 1951 Von Neumann suggested to dad to dad that Economists and Media might be generatively disastrous unless they celebrated questioning future's with engineers. Check out the world Jensen Huang has been inviting humans to linkin since he commited to designing million times more energetic computing including today's AI Chats and deep learning robots.
    India 2024 : 2
    India 2016
    Silicon Valley 2024
    2015 with Elon Musk move video to 97 mins 40 secs
    Valley March 2025.
    Taiwan 2024
    Taiwan XX
    UK Wash DC 2024Japan 2024
    .Is Human Species capable of celebraing intelligence as deeper (and more open) data flow than politicians printing paper money?
    Economistwater.com: Do you know that even the world's biggest nations will fail in 2020s unless their peopled celebrate copiloting waters and energy transmission (CLICK TO PUZZLES of 25% more in 2020s) maps inttrligent;y?
    MOTHER EARTHS CODES: ELERCTRIGICATION POWERS THINGS WITH ELECTRICITY: INTELLIGENCE EMPOWERS PEOPLES: FRESH WATER CONNECTS OUR HEALTH & EMOTIONAL COOL Please linkin with me chris.macrae@yahoo.co.uk (Wash DC) to add where we the peoples can add to these 4 spaces for unearthing humanity's intrlligence boosters-
  • Paris Intelligence Action summit February,
  • Santa Clara future of accelerrated computimng partners- nvidia santa clara Japan's Osaka Expo - 6 months in which any nations pavilion can virally survey intelligence of any other pavilion
  • Canada's G7- will all 7 nations leaders sink or swim together. Of course if we the peoples can decide what inteligences top 20 spaces need to be, we have a chance to change every education momemt og every age of person at every community around the world in line with the future of capitalism that The Economist premiered in 1976.Japan and silicon calley had payed with the first decade of moore's law - would other places be free to entrepremeurially join in the milliion times more compute in time?
  • .Fom Glasgow 1760, engineers artificail system designs became humans 3rd & if integrated respectfully with nature's man's most productive tool alongside bitech waves of water and heat. Long stiory short innovation opportunities and threats to humans in 2025 now that we have 10*18 more tech than 1960s dei=fine surval of commmunities and species everywhere. Ask eg Grok3 whose inteligences or agents 15 year old stidnts and teacgers most need to know about. Benchmark against 6 multipliers
    Jensen Huang or whomever platforms are needed for 2 main types of deep learning ai : bio and making humanoids safe2 Hassabis - whose ai pattern games have made biotech an innovation likely to imact more thahn half of gdp by 2035, often very locally, 100-year catch up with einstein's core maths e=mcsqaured published 1905
    Yann Lecun -we need an academic concerned with open sourcing worldwide student access to what's next - who do you choose if yann lecun not most open academic maths guy inspiring your peoplesMusk - well I am not sure anyone is jealous of musk as biggest brain in the world-don't expect him to suffer jounalist fools lighhtly
    6 Drew endy - or who is sharing with millennials hundreds of startup discoveries of biotech as locally regenerating as batures own source of energy photosyntheis has for millions of years played water and heat games on birthing living systems
    .Lets share some lunchclub updates -more at Economistdiary.com/2020 eg what are varous ways ordinary pcusers are expected to use the new ai pc- at different entry level- for example am i correct that one has to invest about 15000 dollars to be a lamda 3 connector through pc? Or for those (eg parents of 11 year olds in usa 2025-8) who just want eg emails whatsapp zoom (tiktok) linkedin algorithms to sort out 30 years of virtual connections whats min dell system to buy? I am also looking for chapter s[omsors of 2025report.com last of 75 years of mediation by my family and von neumann family and for 40 years The Economist 1951-1990 including now King Charles AI summits- from 1964 Tokyo Olympics win-win between both island post-emires (dad norman awarded honors by both)

    Saturday, September 20, 2025

     Converging human languages on natures maths

    There are thousands of human languages and culturally rich as they often are- we need to have deeeper common sense of how nature's systems work- von neumann saw this as one of main reasons for developing ai's mathematical brainpower

    I think using chats without contexts is at best a waste of time


    lets look at some llms which have not been marketed as chats

    start with nvidia megatron


    AI Overview
    The main use of NVIDIA's Megatron framework has been to train and develop very large-scale language models (LLMs) and other generative AI models
    . The framework was created to overcome the significant computational challenges of training models with billions or trillions of parameters by allowing the work to be split across thousands of GPUs. 
    Core capabilities of Megatron
    Megatron provides a comprehensive, high-performance training library for large transformer models. Its most critical uses include: 
    • Massive model training: The framework uses advanced parallelism strategies—including tensor, pipeline, and sequence parallelism—that enable the training of enormous models like GPT, BERT, and T5 architectures.
    • Distributed training: By efficiently distributing the training workload across multiple GPUs and compute nodes, Megatron enables high-speed training of models that would be computationally impossible on a single machine.
    • Foundation for other frameworks: The techniques pioneered in Megatron have inspired and influenced many other popular LLM development frameworks, such as Hugging Face Accelerate, Colossal-AI, and NVIDIA's own NeMo. 
    Landmark Megatron projects
    The Megatron framework has been instrumental in the development of several notable AI models and systems:
    • Megatron-Turing NLG 530B: A 530-billion-parameter language model developed in collaboration with Microsoft, which pushed the boundaries of natural language generation at the time of its release in 2021.
    • GatorTron: The largest clinical language model, developed by the University of Florida to understand medical texts.
    • BioMegatron: A biomedical language model trained on PubMed and other biomedical datasets to improve performance on specialized tasks.
    • LLM-jp: A large language model with strong Japanese-language capabilities, developed for Japan's Ministry of Economy, Trade and Industry (METI). 
    Applications of Megatron-trained models
    The language models trained using Megatron have been applied to various natural language processing (NLP) tasks, such as:
    • Text generation
    • Question answering
    • Text summarization
    • Machine translation
    • Sentiment analysis
    • Conversational agents and chatbots 
    Megatron has been repurposed for use within NVIDIA's Earth-2 platform, though not in its original form as a large language model (LLM). Instead, NVIDIA leveraged the underlying architecture of Megatron to create a version specifically designed for physics-based AI, called PhysicsNeMo, for large-scale scientific simulations. 
    Here is a breakdown of Megatron's role in NVIDIA's climate and energy projects:
    How Megatron technology is used for Earth-2
    • Physics-informed AI: The Megatron framework, with its massive scaling capabilities, was adapted for NVIDIA's Earth-2 platform, a "digital twin" cloud for simulating climate and weather. The resulting tool, PhysicsNeMo, combines the strengths of large-scale AI with physics-based models to create highly accurate and fast weather and climate simulations.
    • Massive parallel training: Megatron's core innovation was in tensor and pipeline parallelism, which allows training huge models across thousands of GPUs. NVIDIA has extended this to Earth-2 to train models like FourCastNet, which can accelerate weather forecasting with greater energy efficiency than traditional methods.
    • High-resolution simulations: By combining Megatron-based AI with the NVIDIA Modulus framework, Earth-2 can produce high-resolution, kilometer-scale climate simulations. This allows for detailed visualization and prediction of phenomena like storms, floods, and wildfires.
    • Enabling scientific collaboration: The infrastructure, including the Earth-2 platform and its AI tools, has been used in collaborative projects with research institutions like the Max Planck Institute for Meteorology. 
    Key differences between traditional Megatron and its climate application
    It's important to distinguish the specific type of technology involved:
    • Megatron-LM: The original Megatron was a large language model framework, trained on text data to advance the state-of-the-art in LLMs.
    • Earth-2 and PhysicsNeMo: In this context, the name "Megatron" refers to the underlying distributed training techniques used to train large-scale neural network models, which are then applied to climate and weather data rather than text. The models are physics-informed, meaning they are built with the underlying laws of physics incorporated into their design.